insights-operator/gathers.json0000640000000000000000000004353215114274112015256 0ustar0000000000000000{"status_reports":[{"name":"clusterconfig/proxies","duration_in_ms":23,"records_count":1,"errors":null,"warnings":null,"panic":null},{"name":"clusterconfig/openstack_version","duration_in_ms":14,"records_count":0,"errors":null,"warnings":null,"panic":null},{"name":"clusterconfig/sap_pods","duration_in_ms":14,"records_count":0,"errors":null,"warnings":null,"panic":null},{"name":"clusterconfig/ingress","duration_in_ms":20,"records_count":1,"errors":null,"warnings":null,"panic":null},{"name":"clusterconfig/authentication","duration_in_ms":10,"records_count":1,"errors":null,"warnings":null,"panic":null},{"name":"clusterconfig/cost_management_metrics_configs","duration_in_ms":15,"records_count":0,"errors":null,"warnings":null,"panic":null},{"name":"clusterconfig/openshift_logging","duration_in_ms":13,"records_count":0,"errors":null,"warnings":null,"panic":null},{"name":"clusterconfig/active_alerts","duration_in_ms":28,"records_count":0,"errors":["Get \"https://alertmanager-main.openshift-monitoring.svc:9094/api/v2/alerts?active=true\": dial tcp: lookup alertmanager-main.openshift-monitoring.svc on 172.30.0.10:53: no such host"],"warnings":null,"panic":null},{"name":"workloads","duration_in_ms":5134,"records_count":1,"errors":null,"warnings":null,"panic":null},{"name":"clusterconfig/qemu_kubevirt_launcher_logs","duration_in_ms":23,"records_count":0,"errors":null,"warnings":null,"panic":null},{"name":"clusterconfig/openstack_dataplanenodesets","duration_in_ms":13,"records_count":0,"errors":null,"warnings":null,"panic":null},{"name":"clusterconfig/olm_operators","duration_in_ms":22,"records_count":0,"errors":null,"warnings":null,"panic":null},{"name":"clusterconfig/container_images","duration_in_ms":49,"records_count":21,"errors":null,"warnings":null,"panic":null},{"name":"clusterconfig/aggregated_monitoring_cr_names","duration_in_ms":99,"records_count":0,"errors":null,"warnings":null,"panic":null},{"name":"clusterconfig/number_of_pods_and_netnamespaces_with_sdn_annotations","duration_in_ms":62,"records_count":0,"errors":null,"warnings":null,"panic":null},{"name":"clusterconfig/service_accounts","duration_in_ms":12406,"records_count":1,"errors":null,"warnings":null,"panic":null},{"name":"clusterconfig/operators_pods_and_events","duration_in_ms":20065,"records_count":66,"errors":["the record with the same name \"config/pod/openshift-authentication-operator/authentication-operator-6c968fdfdf-nrrfw.json\" was already recorded and had the fingerprint \"c3f15f136e9f1609f6009048bb131df7cedf1c3c36f4cf3a674333449718f5f2\", overwriting with the record having fingerprint \"c3f15f136e9f1609f6009048bb131df7cedf1c3c36f4cf3a674333449718f5f2\"","the record with the same name \"config/pod/openshift-ingress-operator/ingress-operator-8649c48786-cx2b2.json\" was already recorded and had the fingerprint \"15dbc0f5ffbb18f5cd7ed1dd86a316783f8fca281488223800977d32cdd087ab\", overwriting with the record having fingerprint \"15dbc0f5ffbb18f5cd7ed1dd86a316783f8fca281488223800977d32cdd087ab\"","the record with the same name \"config/pod/openshift-ingress/router-default-5465c8b4db-58d52.json\" was already recorded and had the fingerprint \"53664cbe6d8af0b316ba1cdc0f4433a570b4cd69a3849b15c5f25080afd503a5\", overwriting with the record having fingerprint \"53664cbe6d8af0b316ba1cdc0f4433a570b4cd69a3849b15c5f25080afd503a5\"","the record with the same name \"config/pod/openshift-kube-controller-manager/kube-controller-manager-master-0.json\" was already recorded and had the fingerprint \"2ee16ca368a9afa3a3a66635ab78492628f8aa019aa14944568c2a4d80ffc4de\", overwriting with the record having fingerprint \"033b599e40f2fa8787f4abda88343e21e856909b57fa91d8a520f2b500587937\"","the record with the same name \"config/pod/openshift-machine-config-operator/kube-rbac-proxy-crio-master-0.json\" was already recorded and had the fingerprint \"559e3eea240d4e35da871b72504448186d4b04e4e9a4dfccfcd97dfe09e342f4\", overwriting with the record having fingerprint \"559e3eea240d4e35da871b72504448186d4b04e4e9a4dfccfcd97dfe09e342f4\"","the record with the same name \"config/pod/openshift-operator-lifecycle-manager/package-server-manager-67477646d4-7hndf.json\" was already recorded and had the fingerprint \"2c286f025c2166d6b091f22336465017f7a8beb17dbf7361a14da8af52d5ce45\", overwriting with the record having fingerprint \"2c286f025c2166d6b091f22336465017f7a8beb17dbf7361a14da8af52d5ce45\""],"warnings":["warning: the record with the same fingerprint \"c3f15f136e9f1609f6009048bb131df7cedf1c3c36f4cf3a674333449718f5f2\" was already recorded at path \"config/pod/openshift-authentication-operator/authentication-operator-6c968fdfdf-nrrfw.json\", recording another one with a different path \"config/pod/openshift-authentication-operator/authentication-operator-6c968fdfdf-nrrfw.json\"","warning: the record with the same fingerprint \"15dbc0f5ffbb18f5cd7ed1dd86a316783f8fca281488223800977d32cdd087ab\" was already recorded at path \"config/pod/openshift-ingress-operator/ingress-operator-8649c48786-cx2b2.json\", recording another one with a different path \"config/pod/openshift-ingress-operator/ingress-operator-8649c48786-cx2b2.json\"","warning: the record with the same fingerprint \"53664cbe6d8af0b316ba1cdc0f4433a570b4cd69a3849b15c5f25080afd503a5\" was already recorded at path \"config/pod/openshift-ingress/router-default-5465c8b4db-58d52.json\", recording another one with a different path \"config/pod/openshift-ingress/router-default-5465c8b4db-58d52.json\"","warning: the record with the same fingerprint \"8943e5562c2bb6f008439af47e4fcaefa57efc97332190f1e5ea17eb7f6112c1\" was already recorded at path \"config/pod/openshift-kube-controller-manager/logs/kube-controller-manager-master-0/cluster-policy-controller_previous.log\", recording another one with a different path \"config/pod/openshift-kube-controller-manager/logs/kube-controller-manager-master-0/cluster-policy-controller_current.log\"","warning: the record with the same fingerprint \"559e3eea240d4e35da871b72504448186d4b04e4e9a4dfccfcd97dfe09e342f4\" was already recorded at path \"config/pod/openshift-machine-config-operator/kube-rbac-proxy-crio-master-0.json\", recording another one with a different path \"config/pod/openshift-machine-config-operator/kube-rbac-proxy-crio-master-0.json\"","warning: the record with the same fingerprint \"2c286f025c2166d6b091f22336465017f7a8beb17dbf7361a14da8af52d5ce45\" was already recorded at path \"config/pod/openshift-operator-lifecycle-manager/package-server-manager-67477646d4-7hndf.json\", recording another one with a different path \"config/pod/openshift-operator-lifecycle-manager/package-server-manager-67477646d4-7hndf.json\""],"panic":null},{"name":"clusterconfig/sap_datahubs","duration_in_ms":14,"records_count":0,"errors":null,"warnings":null,"panic":null},{"name":"clusterconfig/pdbs","duration_in_ms":22,"records_count":1,"errors":null,"warnings":null,"panic":null},{"name":"clusterconfig/openshift_machine_api_events","duration_in_ms":25,"records_count":1,"errors":null,"warnings":null,"panic":null},{"name":"clusterconfig/overlapping_namespace_uids","duration_in_ms":28,"records_count":1,"errors":["can't read uid range of the openshift-console namespace","can't read uid range of the openshift-console-operator namespace","can't read uid range of the openshift-console-user-settings namespace","can't read uid range of the openshift-ingress-canary namespace"],"warnings":null,"panic":null},{"name":"clusterconfig/version","duration_in_ms":47,"records_count":4,"errors":null,"warnings":null,"panic":null},{"name":"conditional/remote_configuration","duration_in_ms":0,"records_count":1,"errors":null,"warnings":null,"panic":null},{"name":"clusterconfig/machines","duration_in_ms":22,"records_count":0,"errors":null,"warnings":null,"panic":null},{"name":"clusterconfig/machine_autoscalers","duration_in_ms":15,"records_count":0,"errors":["machineautoscalers.autoscaling.openshift.io is forbidden: User \"system:serviceaccount:openshift-insights:gather\" cannot list resource \"machineautoscalers\" in API group \"autoscaling.openshift.io\" at the cluster scope"],"warnings":null,"panic":null},{"name":"clusterconfig/jaegers","duration_in_ms":13,"records_count":0,"errors":null,"warnings":null,"panic":null},{"name":"clusterconfig/nodes","duration_in_ms":23,"records_count":1,"errors":null,"warnings":null,"panic":null},{"name":"clusterconfig/tsdb_status","duration_in_ms":12,"records_count":0,"errors":["Get \"https://prometheus-k8s.openshift-monitoring.svc:9091/api/v1/status/tsdb\": dial tcp: lookup prometheus-k8s.openshift-monitoring.svc on 172.30.0.10:53: no such host"],"warnings":null,"panic":null},{"name":"clusterconfig/infrastructures","duration_in_ms":20,"records_count":1,"errors":null,"warnings":null,"panic":null},{"name":"clusterconfig/storage_classes","duration_in_ms":23,"records_count":0,"errors":null,"warnings":null,"panic":null},{"name":"clusterconfig/validating_webhook_configurations","duration_in_ms":25,"records_count":6,"errors":null,"warnings":null,"panic":null},{"name":"clusterconfig/crds","duration_in_ms":42,"records_count":2,"errors":null,"warnings":null,"panic":null},{"name":"clusterconfig/clusterroles","duration_in_ms":29,"records_count":2,"errors":null,"warnings":null,"panic":null},{"name":"clusterconfig/operators","duration_in_ms":79,"records_count":34,"errors":null,"warnings":null,"panic":null},{"name":"clusterconfig/machine_config_pools","duration_in_ms":24,"records_count":2,"errors":null,"warnings":null,"panic":null},{"name":"clusterconfig/nodenetworkconfigurationpolicies","duration_in_ms":24,"records_count":0,"errors":null,"warnings":null,"panic":null},{"name":"clusterconfig/storage_cluster","duration_in_ms":8,"records_count":0,"errors":null,"warnings":null,"panic":null},{"name":"clusterconfig/oauths","duration_in_ms":10,"records_count":1,"errors":null,"warnings":null,"panic":null},{"name":"clusterconfig/certificate_signing_requests","duration_in_ms":15,"records_count":0,"errors":null,"warnings":null,"panic":null},{"name":"clusterconfig/schedulers","duration_in_ms":24,"records_count":1,"errors":null,"warnings":null,"panic":null},{"name":"clusterconfig/machine_healthchecks","duration_in_ms":25,"records_count":1,"errors":null,"warnings":null,"panic":null},{"name":"clusterconfig/lokistack","duration_in_ms":15,"records_count":0,"errors":null,"warnings":null,"panic":null},{"name":"workloads/helmchart_info","duration_in_ms":24,"records_count":0,"errors":null,"warnings":null,"panic":null},{"name":"conditional/conditional_gatherer_rules","duration_in_ms":0,"records_count":1,"errors":null,"warnings":null,"panic":null},{"name":"clusterconfig/cluster_apiserver","duration_in_ms":22,"records_count":1,"errors":null,"warnings":null,"panic":null},{"name":"clusterconfig/node_logs","duration_in_ms":178,"records_count":1,"errors":null,"warnings":null,"panic":null},{"name":"clusterconfig","duration_in_ms":20086,"records_count":169,"errors":["function \"support_secret\" failed with an error","function \"machine_autoscalers\" failed with an error","function \"pod_network_connectivity_checks\" failed with an error","function \"tsdb_status\" failed with an error","function \"metrics\" failed with an error","function \"silenced_alerts\" failed with an error","function \"active_alerts\" failed with an error","function \"overlapping_namespace_uids\" failed with an error","function \"overlapping_namespace_uids\" failed with an error","function \"overlapping_namespace_uids\" failed with an error","function \"overlapping_namespace_uids\" failed with an error","function \"config_maps\" failed with an error","function \"config_maps\" failed with an error","function \"config_maps\" failed with an error","unable to record function \"operators_pods_and_events\" record \"config/pod/openshift-authentication-operator/authentication-operator-6c968fdfdf-nrrfw.json\"","unable to record function \"operators_pods_and_events\" record \"config/pod/openshift-ingress-operator/ingress-operator-8649c48786-cx2b2.json\"","unable to record function \"operators_pods_and_events\" record \"config/pod/openshift-ingress/router-default-5465c8b4db-58d52.json\"","unable to record function \"operators_pods_and_events\" record \"config/pod/openshift-kube-controller-manager/kube-controller-manager-master-0.json\"","unable to record function \"operators_pods_and_events\" record \"config/pod/openshift-machine-config-operator/kube-rbac-proxy-crio-master-0.json\"","unable to record function \"operators_pods_and_events\" record \"config/pod/openshift-operator-lifecycle-manager/package-server-manager-67477646d4-7hndf.json\""],"warnings":null,"panic":null},{"name":"clusterconfig/container_runtime_configs","duration_in_ms":24,"records_count":0,"errors":null,"warnings":null,"panic":null},{"name":"clusterconfig/networks","duration_in_ms":26,"records_count":1,"errors":null,"warnings":null,"panic":null},{"name":"clusterconfig/pod_network_connectivity_checks","duration_in_ms":22,"records_count":0,"errors":["the server could not find the requested resource (get podnetworkconnectivitychecks.controlplane.operator.openshift.io)"],"warnings":null,"panic":null},{"name":"clusterconfig/openstack_dataplanedeployments","duration_in_ms":12,"records_count":0,"errors":null,"warnings":null,"panic":null},{"name":"clusterconfig/mutating_webhook_configurations","duration_in_ms":22,"records_count":1,"errors":null,"warnings":null,"panic":null},{"name":"conditional/rapid_container_logs","duration_in_ms":0,"records_count":0,"errors":null,"warnings":null,"panic":null},{"name":"workloads/workload_info","duration_in_ms":5134,"records_count":1,"errors":null,"warnings":null,"panic":null},{"name":"clusterconfig/support_secret","duration_in_ms":24,"records_count":0,"errors":["secrets \"support\" not found"],"warnings":null,"panic":null},{"name":"clusterconfig/image_registries","duration_in_ms":15,"records_count":1,"errors":null,"warnings":null,"panic":null},{"name":"clusterconfig/silenced_alerts","duration_in_ms":16,"records_count":0,"errors":["Get \"https://alertmanager-main.openshift-monitoring.svc:9094/api/v2/alerts?active=false\u0026inhibited=false\u0026silenced=true\": dial tcp: lookup alertmanager-main.openshift-monitoring.svc on 172.30.0.10:53: no such host"],"warnings":null,"panic":null},{"name":"clusterconfig/machine_configs","duration_in_ms":50,"records_count":1,"errors":null,"warnings":null,"panic":null},{"name":"clusterconfig/config_maps","duration_in_ms":51,"records_count":9,"errors":["configmaps \"cluster-monitoring-config\" not found","configmaps \"gateway-mode-config\" not found","configmaps \"insights-config\" not found"],"warnings":["warning: the record with the same fingerprint \"34cdf5d36d45a9e3ee7c2be31546e3d937ae4ebf144ac07b0210c06a1610a94d\" was already recorded at path \"config/configmaps/openshift-config/admin-kubeconfig-client-ca/ca-bundle.crt\", recording another one with a different path \"config/configmaps/openshift-config/etcd-ca-bundle/ca-bundle.crt\"","warning: the record with the same fingerprint \"34cdf5d36d45a9e3ee7c2be31546e3d937ae4ebf144ac07b0210c06a1610a94d\" was already recorded at path \"config/configmaps/openshift-config/etcd-ca-bundle/ca-bundle.crt\", recording another one with a different path \"config/configmaps/openshift-config/etcd-serving-ca/ca-bundle.crt\"","warning: the record with the same fingerprint \"34cdf5d36d45a9e3ee7c2be31546e3d937ae4ebf144ac07b0210c06a1610a94d\" was already recorded at path \"config/configmaps/openshift-config/etcd-serving-ca/ca-bundle.crt\", recording another one with a different path \"config/configmaps/openshift-config/openshift-service-ca.crt/service-ca.crt\""],"panic":null},{"name":"clusterconfig/metrics","duration_in_ms":26,"records_count":0,"errors":["Get \"https://prometheus-k8s.openshift-monitoring.svc:9091/federate?match%5B%5D=cluster_installer\u0026match%5B%5D=namespace%3Acontainer_cpu_usage%3Asum\u0026match%5B%5D=namespace%3Acontainer_memory_usage_bytes%3Asum\u0026match%5B%5D=vsphere_node_hw_version_total\u0026match%5B%5D=virt_platform\u0026match%5B%5D=console_helm_installs_total\u0026match%5B%5D=console_helm_upgrades_total\u0026match%5B%5D=console_helm_uninstalls_total\u0026match%5B%5D=openshift_apps_deploymentconfigs_strategy_total\u0026match%5B%5D=etcd_server_slow_apply_total\u0026match%5B%5D=etcd_server_slow_read_indexes_total\u0026match%5B%5D=haproxy_exporter_server_threshold\": dial tcp: lookup prometheus-k8s.openshift-monitoring.svc on 172.30.0.10:53: no such host"],"warnings":null,"panic":null},{"name":"clusterconfig/install_plans","duration_in_ms":11641,"records_count":1,"errors":null,"warnings":null,"panic":null},{"name":"clusterconfig/nodenetworkstates","duration_in_ms":16,"records_count":0,"errors":null,"warnings":null,"panic":null},{"name":"clusterconfig/openstack_controlplanes","duration_in_ms":24,"records_count":0,"errors":null,"warnings":null,"panic":null},{"name":"clusterconfig/image","duration_in_ms":22,"records_count":1,"errors":null,"warnings":null,"panic":null},{"name":"clusterconfig/ceph_cluster","duration_in_ms":15,"records_count":0,"errors":null,"warnings":null,"panic":null},{"name":"clusterconfig/sap_config","duration_in_ms":10,"records_count":0,"errors":null,"warnings":null,"panic":null},{"name":"clusterconfig/monitoring_persistent_volumes","duration_in_ms":19,"records_count":0,"errors":null,"warnings":null,"panic":null},{"name":"clusterconfig/image_pruners","duration_in_ms":22,"records_count":1,"errors":null,"warnings":null,"panic":null},{"name":"conditional","duration_in_ms":296,"records_count":2,"errors":null,"warnings":null,"panic":null},{"name":"clusterconfig/feature_gates","duration_in_ms":23,"records_count":1,"errors":null,"warnings":null,"panic":null},{"name":"clusterconfig/dvo_metrics","duration_in_ms":23,"records_count":0,"errors":null,"warnings":["warning: no service found with label selector name=deployment-validation-operator"],"panic":null},{"name":"clusterconfig/machine_sets","duration_in_ms":13,"records_count":0,"errors":null,"warnings":null,"panic":null},{"name":"clusterconfig/ingress_certificates","duration_in_ms":37,"records_count":1,"errors":null,"warnings":null,"panic":null}],"container_memory_bytes_usage":0,"uptime_seconds":1035.586,"is_global_obfuscation_enabled":false}insights-operator/remote-configuration.json0000640000000000000000000000443515114274112017760 0ustar0000000000000000{"conditional_gathering_rules":[{"conditions":[{"alert":{"name":"APIRemovedInNextEUSReleaseInUse"},"type":"alert_is_firing"}],"gathering_functions":{"api_request_counts_of_resource_from_alert":{"alert_name":"APIRemovedInNextEUSReleaseInUse"}}},{"conditions":[{"alert":{"name":"AlertmanagerFailedReload"},"type":"alert_is_firing"}],"gathering_functions":{"containers_logs":{"alert_name":"AlertmanagerFailedReload","container":"alertmanager","tail_lines":50}}},{"conditions":[{"alert":{"name":"AlertmanagerFailedToSendAlerts"},"type":"alert_is_firing"}],"gathering_functions":{"containers_logs":{"alert_name":"AlertmanagerFailedToSendAlerts","container":"alertmanager","tail_lines":50}}},{"conditions":[{"alert":{"name":"KubePodCrashLooping"},"type":"alert_is_firing"}],"gathering_functions":{"containers_logs":{"alert_name":"KubePodCrashLooping","previous":true,"tail_lines":20}}},{"conditions":[{"alert":{"name":"KubePodNotReady"},"type":"alert_is_firing"}],"gathering_functions":{"containers_logs":{"alert_name":"KubePodNotReady","tail_lines":100},"pod_definition":{"alert_name":"KubePodNotReady"}}},{"conditions":[{"alert":{"name":"PrometheusOperatorSyncFailed"},"type":"alert_is_firing"}],"gathering_functions":{"containers_logs":{"alert_name":"PrometheusOperatorSyncFailed","container":"prometheus-operator","tail_lines":50}}},{"conditions":[{"alert":{"name":"PrometheusRemoteStorageFailures"},"type":"alert_is_firing"}],"gathering_functions":{"containers_logs":{"alert_name":"PrometheusRemoteStorageFailures","container":"prometheus","tail_lines":100}}},{"conditions":[{"alert":{"name":"PrometheusTargetSyncFailure"},"type":"alert_is_firing"}],"gathering_functions":{"containers_logs":{"alert_name":"PrometheusTargetSyncFailure","container":"prometheus","tail_lines":50}}},{"conditions":[{"alert":{"name":"SamplesImagestreamImportFailing"},"type":"alert_is_firing"}],"gathering_functions":{"image_streams_of_namespace":{"namespace":"openshift-cluster-samples-operator"},"logs_of_namespace":{"namespace":"openshift-cluster-samples-operator","tail_lines":100}}},{"conditions":[{"alert":{"name":"ThanosRuleQueueIsDroppingAlerts"},"type":"alert_is_firing"}],"gathering_functions":{"containers_logs":{"alert_name":"ThanosRuleQueueIsDroppingAlerts","container":"thanos-ruler","tail_lines":50}}}],"container_logs":[],"version":"1.2.0"}insights-operator/conditional-gatherer-rules.json0000640000000000000000000000577515114274112021062 0ustar0000000000000000{"version":"1.2.0","conditional_gathering_rules":[{"rule":{"conditions":[{"type":"alert_is_firing","alert":{"name":"APIRemovedInNextEUSReleaseInUse"}}],"gathering_functions":{"api_request_counts_of_resource_from_alert":{"alert_name":"APIRemovedInNextEUSReleaseInUse"}}},"errors":["alerts cache is missing"],"was_triggered":false},{"rule":{"conditions":[{"type":"alert_is_firing","alert":{"name":"AlertmanagerFailedReload"}}],"gathering_functions":{"containers_logs":{"alert_name":"AlertmanagerFailedReload","container":"alertmanager","tail_lines":50}}},"errors":["alerts cache is missing"],"was_triggered":false},{"rule":{"conditions":[{"type":"alert_is_firing","alert":{"name":"AlertmanagerFailedToSendAlerts"}}],"gathering_functions":{"containers_logs":{"alert_name":"AlertmanagerFailedToSendAlerts","container":"alertmanager","tail_lines":50}}},"errors":["alerts cache is missing"],"was_triggered":false},{"rule":{"conditions":[{"type":"alert_is_firing","alert":{"name":"KubePodCrashLooping"}}],"gathering_functions":{"containers_logs":{"alert_name":"KubePodCrashLooping","tail_lines":20,"previous":true}}},"errors":["alerts cache is missing"],"was_triggered":false},{"rule":{"conditions":[{"type":"alert_is_firing","alert":{"name":"KubePodNotReady"}}],"gathering_functions":{"containers_logs":{"alert_name":"KubePodNotReady","tail_lines":100},"pod_definition":{"alert_name":"KubePodNotReady"}}},"errors":["alerts cache is missing"],"was_triggered":false},{"rule":{"conditions":[{"type":"alert_is_firing","alert":{"name":"PrometheusOperatorSyncFailed"}}],"gathering_functions":{"containers_logs":{"alert_name":"PrometheusOperatorSyncFailed","container":"prometheus-operator","tail_lines":50}}},"errors":["alerts cache is missing"],"was_triggered":false},{"rule":{"conditions":[{"type":"alert_is_firing","alert":{"name":"PrometheusRemoteStorageFailures"}}],"gathering_functions":{"containers_logs":{"alert_name":"PrometheusRemoteStorageFailures","container":"prometheus","tail_lines":100}}},"errors":["alerts cache is missing"],"was_triggered":false},{"rule":{"conditions":[{"type":"alert_is_firing","alert":{"name":"PrometheusTargetSyncFailure"}}],"gathering_functions":{"containers_logs":{"alert_name":"PrometheusTargetSyncFailure","container":"prometheus","tail_lines":50}}},"errors":["alerts cache is missing"],"was_triggered":false},{"rule":{"conditions":[{"type":"alert_is_firing","alert":{"name":"SamplesImagestreamImportFailing"}}],"gathering_functions":{"image_streams_of_namespace":{"namespace":"openshift-cluster-samples-operator"},"logs_of_namespace":{"namespace":"openshift-cluster-samples-operator","tail_lines":100}}},"errors":["alerts cache is missing"],"was_triggered":false},{"rule":{"conditions":[{"type":"alert_is_firing","alert":{"name":"ThanosRuleQueueIsDroppingAlerts"}}],"gathering_functions":{"containers_logs":{"alert_name":"ThanosRuleQueueIsDroppingAlerts","container":"thanos-ruler","tail_lines":50}}},"errors":["alerts cache is missing"],"was_triggered":false}],"endpoint":"https://console.redhat.com/api/gathering/v2/%s/gathering_rules"}config/workload_info.json0000640000000000000000000005306315114274112014242 0ustar0000000000000000{"pods":76,"imageCount":61,"images":{"sha256:79f99fd6cce984287932edf0d009660bb488d663081f3d62ec3b23bc8bfbf6c2":{"layerIDs":["sha256:25c75c34b2e2b68ba9245d9cddeb6b8a0887371ed30744064f85241a75704d87","sha256:a7afe321285e421a0de5ee7eb5b7f1872a31252aedbece4162ec13a35b1c88e0","sha256:b45b4080e75db66dbb2f4d8403f29133c1829a6e7a5055752f4267aea3a23894","sha256:52445dcc2525a6a1af5c23690894dfb632a19a8912d9b564ac304978a4fe2d77"],"repository":"vXIOZyWw1exO"}},"namespaces":{"0LiT6ZNtbpYL":{"count":4,"shapes":[{"restartAlways":true,"initContainers":[{"imageID":"sha256:cfde59e48cd5dee3721f34d249cb119cc3259fd857965d34f9c7ed83b0c363a1","firstCommand":"Cl6kTzfbYztA"},{"imageID":"sha256:916566bb9d0143352324233d460ad94697719c11c8c9158e3aea8f475941751f","firstCommand":"N9KxLV2avCo2"},{"imageID":"sha256:8a3d37aa7a22c68afa963ecfb4b43c52cccf152580cd66e4d5382fb69e4037cc","firstCommand":"Cl6kTzfbYztA"},{"imageID":"sha256:9432c13d76bd4ba4eb9197c050cf88c0d701fa2055eeb59257e2e23901f9fdff","firstCommand":"Cl6kTzfbYztA"},{"imageID":"sha256:631a3798b749fecc041a99929eb946618df723e15055e805ff752a1a1273481c","firstCommand":"Cl6kTzfbYztA"},{"imageID":"sha256:631a3798b749fecc041a99929eb946618df723e15055e805ff752a1a1273481c","firstCommand":"icTsn2s_EIax"}],"containers":[{"imageID":"sha256:9014f384de5f9a0b7418d5869ad349abb9588d16bd09ed650a163c045315dbff","firstCommand":"N9KxLV2avCo2","firstArg":"EbplhSJxzSTF"}]},{"restartAlways":true,"containers":[{"imageID":"sha256:c69ffd8f8dcceedc2d6eb306cea33f8beabc1be1308cd5f4ee8b9a8e3eab9843","firstArg":"47DEQpj8HBSa"},{"imageID":"sha256:4ecc5bac651ff1942865baee5159582e9602c89b47eeab18400a32abcba8f690","firstCommand":"N9KxLV2avCo2"}]},{"restartAlways":true,"containers":[{"imageID":"sha256:9014f384de5f9a0b7418d5869ad349abb9588d16bd09ed650a163c045315dbff","firstCommand":"N9KxLV2avCo2","firstArg":"BuLIUMMJnyP_"}]},{"restartAlways":true,"containers":[{"imageID":"sha256:c69ffd8f8dcceedc2d6eb306cea33f8beabc1be1308cd5f4ee8b9a8e3eab9843","firstArg":"47DEQpj8HBSa"},{"imageID":"sha256:2632d7f05d5a992e91038ded81c715898f3fe803420a9b67a0201e9fd8075213","firstCommand":"FC0eu9ID58Bm","firstArg":"47DEQpj8HBSa"}]}]},"0h5ohwdJx4S5":{"count":1,"shapes":[{"restartAlways":true,"containers":[{"imageID":"sha256:c69ffd8f8dcceedc2d6eb306cea33f8beabc1be1308cd5f4ee8b9a8e3eab9843","firstArg":"aV86_fLS8Aye"},{"imageID":"sha256:1f952cec1e5332b84bdffa249cd426f39087058d6544ddcec650a414c15a9b68","firstCommand":"buSkac1OkQU4","firstArg":"XV1E-c-gSQmZ"}]}]},"2IqVYeidVUxV":{"count":1,"shapes":[{"restartAlways":true,"initContainers":[{"imageID":"sha256:b00c658332d6c6786bd969b26097c20a78c79c045f1692a8809234f5fb586c22","firstCommand":"po5yju59c7HJ"}],"containers":[{"imageID":"sha256:3b8d91a25eeb9f02041e947adb3487da3e7ab8449d3d2ad015827e7954df7b34","firstCommand":"UcO4la9rqCTe"}]}]},"43Xz0HbE9FHX":{"count":5,"shapes":[{"restartAlways":true,"initContainers":[{"imageID":"sha256:f1ca78c423f43f89a0411e40393642f64e4f8df9e5f61c25e31047c4cce170f9","firstCommand":"5EtL1wfFQLyo","firstArg":"_lK3daI6CUb2"},{"imageID":"sha256:98141b7d0b74195f511ccd4c41ba6c8a69fef01f22b2fb8ef2d024109e93a3a9","firstCommand":"_lK3daI6CUb2","firstArg":"1CqC9ME648jS"}],"containers":[{"imageID":"sha256:01d2e67fd74086da701c39dac5b821822351cb0151f9afe72821c05df19953ad","firstCommand":"VDU_nFNG7OR_","firstArg":"JMRYz7RtmkVt"}]},{"restartAlways":true,"initContainers":[{"imageID":"sha256:f1ca78c423f43f89a0411e40393642f64e4f8df9e5f61c25e31047c4cce170f9","firstCommand":"5EtL1wfFQLyo","firstArg":"_lK3daI6CUb2"},{"imageID":"sha256:9eadf94ba5e517edf53c9b21d83f18cb3790ddc768f10d0ca92be97c56a0eecb","firstCommand":"_lK3daI6CUb2","firstArg":"1CqC9ME648jS"}],"containers":[{"imageID":"sha256:01d2e67fd74086da701c39dac5b821822351cb0151f9afe72821c05df19953ad","firstCommand":"VDU_nFNG7OR_","firstArg":"JMRYz7RtmkVt"}]},{"restartAlways":true,"containers":[{"imageID":"sha256:7664a2d4cb10e82ed32abbf95799f43fc3d10135d7dd94799730de504a89680a","firstCommand":"4MEQ8E7-fUvu","firstArg":"mDEjdpyJDBZ4"}]},{"restartAlways":true,"initContainers":[{"imageID":"sha256:f1ca78c423f43f89a0411e40393642f64e4f8df9e5f61c25e31047c4cce170f9","firstCommand":"5EtL1wfFQLyo","firstArg":"_lK3daI6CUb2"},{"imageID":"sha256:e8990432556acad31519b1a73ec32f32d27c2034cf9e5cc4db8980efc7331594","firstCommand":"_lK3daI6CUb2","firstArg":"1CqC9ME648jS"}],"containers":[{"imageID":"sha256:01d2e67fd74086da701c39dac5b821822351cb0151f9afe72821c05df19953ad","firstCommand":"VDU_nFNG7OR_","firstArg":"JMRYz7RtmkVt"}]},{"restartAlways":true,"initContainers":[{"imageID":"sha256:f1ca78c423f43f89a0411e40393642f64e4f8df9e5f61c25e31047c4cce170f9","firstCommand":"5EtL1wfFQLyo","firstArg":"_lK3daI6CUb2"},{"imageID":"sha256:2efbb545a141552851226bea008b13d92cbb084339bcfd6923b38d23c382145e","firstCommand":"_lK3daI6CUb2","firstArg":"1CqC9ME648jS"}],"containers":[{"imageID":"sha256:01d2e67fd74086da701c39dac5b821822351cb0151f9afe72821c05df19953ad","firstCommand":"VDU_nFNG7OR_","firstArg":"JMRYz7RtmkVt"}]}]},"47HAKWqp7-jn":{"count":3,"terminalCount":2,"shapes":[{"restartAlways":true,"initContainers":[{"imageID":"sha256:188637a52cafee61ec461e92fb0c605e28be325b9ac1f2ac8a37d68e97654718","firstCommand":"icTsn2s_EIax"},{"imageID":"sha256:188637a52cafee61ec461e92fb0c605e28be325b9ac1f2ac8a37d68e97654718","firstCommand":"icTsn2s_EIax"},{"imageID":"sha256:188637a52cafee61ec461e92fb0c605e28be325b9ac1f2ac8a37d68e97654718","firstCommand":"icTsn2s_EIax"}],"containers":[{"imageID":"sha256:188637a52cafee61ec461e92fb0c605e28be325b9ac1f2ac8a37d68e97654718","firstCommand":"icTsn2s_EIax"},{"imageID":"sha256:188637a52cafee61ec461e92fb0c605e28be325b9ac1f2ac8a37d68e97654718","firstCommand":"icTsn2s_EIax"},{"imageID":"sha256:f4d4282cb53325e737ad68abbfcb70687ae04fb50353f4f0ba0ba5703b15009a","firstCommand":"icTsn2s_EIax"},{"imageID":"sha256:f4d4282cb53325e737ad68abbfcb70687ae04fb50353f4f0ba0ba5703b15009a","firstCommand":"icTsn2s_EIax"},{"imageID":"sha256:188637a52cafee61ec461e92fb0c605e28be325b9ac1f2ac8a37d68e97654718","firstCommand":"N9KxLV2avCo2"}]}]},"4k7XyKW2A17e":{"count":2,"terminalCount":1,"shapes":[{"restartAlways":true,"initContainers":[{"imageID":"sha256:6d5891cdd7dcf7c9081de8b364b4c96446b7f946f7880fbae291a4592a198264","firstCommand":"930btY2ohuPL","firstArg":"CSx56PgOVZ5A"}],"containers":[{"imageID":"sha256:6d5891cdd7dcf7c9081de8b364b4c96446b7f946f7880fbae291a4592a198264","firstCommand":"aC8iSHmKHojh","firstArg":"r7zEsGNvlNrk"},{"imageID":"sha256:f042fa25014f3d37f3ea967d21f361d2a11833ae18f2c750318101b25d2497ce","firstCommand":"WFJEBSInAJpw","firstArg":"5gS4z5mOnDuY"},{"imageID":"sha256:f042fa25014f3d37f3ea967d21f361d2a11833ae18f2c750318101b25d2497ce","firstCommand":"N9KxLV2avCo2","firstArg":"930btY2ohuPL"}]}]},"5fn-sCitjq6a":{"count":3,"shapes":[{"restartAlways":true,"containers":[{"imageID":"sha256:97d26892192b552c16527bf2771e1b86528ab581a02dd9279cdf71c194830e3e","firstCommand":"HgS-rIunFHfD","firstArg":"zO0oxtw_mcI5"}]},{"restartAlways":true,"containers":[{"imageID":"sha256:d3ce2cbf1032ad0f24f204db73687002fcf302e86ebde3945801c74351b64576","firstArg":"ESRPWfqdsWE1"}]},{"restartAlways":true,"containers":[{"imageID":"sha256:10e57ca7611f79710f05777dc6a8f31c7e04eb09da4d8d793a5acfbf0e4692d7","firstArg":"zO0oxtw_mcI5"}]}]},"6KbiDGNUd_ub":{"count":1,"shapes":[{"restartAlways":true,"containers":[{"imageID":"sha256:e3e65409fc2b27ad0aaeb500a39e264663d2980821f099b830b551785ce4ce8b","firstCommand":"N9KxLV2avCo2"},{"imageID":"sha256:e3e65409fc2b27ad0aaeb500a39e264663d2980821f099b830b551785ce4ce8b","firstCommand":"N9KxLV2avCo2"}]}]},"6bLpdRMvgXVP":{"count":1,"shapes":[{"restartAlways":true,"containers":[{"imageID":"sha256:7e438b814f8e16f00b3fc4b69991af80eee79ae111d2a707f34aa64b2ccbb6eb","firstCommand":"N9KxLV2avCo2","firstArg":"EbplhSJxzSTF"},{"imageID":"sha256:7e438b814f8e16f00b3fc4b69991af80eee79ae111d2a707f34aa64b2ccbb6eb","firstCommand":"N9KxLV2avCo2","firstArg":"EbplhSJxzSTF"}]}]},"9dJB3pvMrhr8":{"count":1,"shapes":[{"restartAlways":true,"containers":[{"imageID":"sha256:c416b201d480bddb5a4960ec42f4740761a1335001cf84ba5ae19ad6857771b1","firstCommand":"51QuNwyM-xYf","firstArg":"r7zEsGNvlNrk"}]}]},"AN3N38xANElq":{"count":3,"terminalCount":2,"shapes":[{"restartAlways":true,"containers":[{"imageID":"sha256:d64c13fe7663a0b4ae61d103b1b7598adcf317a01826f296bcb66b1a2de83c96","firstCommand":"N9KxLV2avCo2","firstArg":"930btY2ohuPL"},{"imageID":"sha256:6d5891cdd7dcf7c9081de8b364b4c96446b7f946f7880fbae291a4592a198264","firstCommand":"N9KxLV2avCo2","firstArg":"930btY2ohuPL"},{"imageID":"sha256:a824e468cf8dd61d347e35b2ee5bc2f815666957647098e21a1bb56ff613e5b9","firstCommand":"FATPL2W-sitV","firstArg":"5gS4z5mOnDuY"},{"imageID":"sha256:a824e468cf8dd61d347e35b2ee5bc2f815666957647098e21a1bb56ff613e5b9","firstCommand":"N9KxLV2avCo2","firstArg":"930btY2ohuPL"}]}]},"CMkFy_B6OMUO":{"count":1,"shapes":[{"restartAlways":true,"containers":[{"imageID":"sha256:8139ed65c0a0a4b0f253b715c11cc52be027efe8a4774da9ccce35c78ef439da","firstCommand":"7aQQEh6Tr2m7","firstArg":"r7zEsGNvlNrk"}]}]},"C_9an18W0cU6":{"count":4,"shapes":[{"restartAlways":true,"containers":[{"imageID":"sha256:f1ca78c423f43f89a0411e40393642f64e4f8df9e5f61c25e31047c4cce170f9","firstCommand":"ZS9VAWJDvxuf","firstArg":"47DEQpj8HBSa"}]},{"restartAlways":true,"containers":[{"imageID":"sha256:f1ca78c423f43f89a0411e40393642f64e4f8df9e5f61c25e31047c4cce170f9","firstCommand":"ArvlNzaZtYpC","firstArg":"47DEQpj8HBSa"}]},{"restartAlways":true,"containers":[{"imageID":"sha256:c69ffd8f8dcceedc2d6eb306cea33f8beabc1be1308cd5f4ee8b9a8e3eab9843","firstArg":"aV86_fLS8Aye"},{"imageID":"sha256:f1ca78c423f43f89a0411e40393642f64e4f8df9e5f61c25e31047c4cce170f9","firstCommand":"PPJ_dAh1Wl4S","firstArg":"47DEQpj8HBSa"}]},{"restartAlways":true,"containers":[{"imageID":"sha256:f1ca78c423f43f89a0411e40393642f64e4f8df9e5f61c25e31047c4cce170f9","firstCommand":"5-ZY-7KtGiOe"}]}]},"EOB8Fqt74gVH":{"count":1,"shapes":[{"restartAlways":true,"containers":[{"imageID":"sha256:61664aa69b33349cc6de45e44ae6033e7f483c034ea01c0d9a8ca08a12d88e3a","firstCommand":"N9KxLV2avCo2","firstArg":"k19oMZ1PIn4C"},{"imageID":"sha256:c69ffd8f8dcceedc2d6eb306cea33f8beabc1be1308cd5f4ee8b9a8e3eab9843","firstArg":"aV86_fLS8Aye"}]}]},"EyQn00UhCg7z":{"count":2,"ignoredCount":1,"shapes":[{"restartAlways":true,"containers":[{"imageID":"sha256:9724d2036305cbd729e1f484c5bad89971de977fff8a6723fef1873858dd1123","firstCommand":"ZDrZSWYe4Hpn"}]}]},"F-y2qosZEzxd":{"count":1,"shapes":[{"restartAlways":true,"containers":[{"imageID":"sha256:8139ed65c0a0a4b0f253b715c11cc52be027efe8a4774da9ccce35c78ef439da","firstCommand":"7aQQEh6Tr2m7","firstArg":"gcNszUTvGLqr"}]}]},"ICyEPw5KVYuy":{"count":2,"shapes":[{"restartAlways":true,"containers":[{"imageID":"sha256:eb928c13a46d3fb45f4a881892d023a92d610a5430be0ffd916aaf8da8e7d297","firstCommand":"ijLlFtdBW-Uk","firstArg":"47DEQpj8HBSa"},{"imageID":"sha256:c69ffd8f8dcceedc2d6eb306cea33f8beabc1be1308cd5f4ee8b9a8e3eab9843","firstArg":"47DEQpj8HBSa"}]},{"restartAlways":true,"containers":[{"imageID":"sha256:79f99fd6cce984287932edf0d009660bb488d663081f3d62ec3b23bc8bfbf6c2","firstCommand":"N9KxLV2avCo2"}]}]},"KETRwZOPBpF9":{"count":1,"shapes":[{"restartAlways":true,"containers":[{"imageID":"sha256:c69ffd8f8dcceedc2d6eb306cea33f8beabc1be1308cd5f4ee8b9a8e3eab9843","firstArg":"aV86_fLS8Aye"},{"imageID":"sha256:3f0aa9cd04713acc5c6fea721bd849e1500da8ae945e0b32000887f34d786e0b","firstCommand":"buSkac1OkQU4","firstArg":"47DEQpj8HBSa"}]}]},"LBp1kjMRhgCD":{"count":1,"shapes":[{"restartAlways":true,"initContainers":[{"imageID":"sha256:91af633e585621630c40d14f188e37d36b44678d0a59e582d850bf8d593d3a0c","firstCommand":"icTsn2s_EIax"}],"containers":[{"imageID":"sha256:91af633e585621630c40d14f188e37d36b44678d0a59e582d850bf8d593d3a0c","firstCommand":"N9KxLV2avCo2","firstArg":"k19oMZ1PIn4C"}]}]},"MBms5agiSYKb":{"count":1,"shapes":[{"restartAlways":true,"containers":[{"imageID":"sha256:33a20002692769235e95271ab071783c57ff50681088fa1035b86af31e73cf20","firstArg":"zO0oxtw_mcI5"}]}]},"N3z7Gdl4s2xN":{"count":1,"shapes":[{"restartAlways":true,"containers":[{"imageID":"sha256:3c1edf52f70bf9b1d1457e0c4111bc79cdaa1edd659ddbdb9d8176eff8b46956","firstCommand":"vLhjmt4GJVrQ"},{"imageID":"sha256:c69ffd8f8dcceedc2d6eb306cea33f8beabc1be1308cd5f4ee8b9a8e3eab9843","firstArg":"47DEQpj8HBSa"}]}]},"NJ5StnkN7lNJ":{"count":2,"shapes":[{"restartAlways":true,"containers":[{"imageID":"sha256:5451aa441e5b8d8689c032405d410c8049a849ef2edf77e5b6a5ce2838c6569b","firstCommand":"euUHIZmT1r6w","firstArg":"gcNszUTvGLqr"}]},{"restartAlways":true,"containers":[{"imageID":"sha256:5451aa441e5b8d8689c032405d410c8049a849ef2edf77e5b6a5ce2838c6569b","firstCommand":"euUHIZmT1r6w"}]}]},"NuNApgGLWU3S":{"count":1,"shapes":[{"restartAlways":true,"containers":[{"imageID":"sha256:1386b0fcb731d843f15fb64532f8b676c927821d69dd3d4503c973c3e2a04216","firstCommand":"loDbMsNsNHAN"},{"imageID":"sha256:1386b0fcb731d843f15fb64532f8b676c927821d69dd3d4503c973c3e2a04216","firstCommand":"BWdk-C4p4WVi","firstArg":"ov8ZFLf4OaCu"}]}]},"QRCRK87Nh9-g":{"count":3,"terminalCount":2,"shapes":[{"restartAlways":true,"initContainers":[{"imageID":"sha256:6d5891cdd7dcf7c9081de8b364b4c96446b7f946f7880fbae291a4592a198264","firstCommand":"930btY2ohuPL","firstArg":"CSx56PgOVZ5A"}],"containers":[{"imageID":"sha256:6d5891cdd7dcf7c9081de8b364b4c96446b7f946f7880fbae291a4592a198264","firstCommand":"N9KxLV2avCo2","firstArg":"OOfnAGgDw729"},{"imageID":"sha256:0f43c31aa3359159d4557dad3cfaf812d8ce44db9cb9ae970e06d3479070b660","firstCommand":"VvlllczXFX2z","firstArg":"5gS4z5mOnDuY"},{"imageID":"sha256:0f43c31aa3359159d4557dad3cfaf812d8ce44db9cb9ae970e06d3479070b660","firstCommand":"VvlllczXFX2z","firstArg":"5gS4z5mOnDuY"},{"imageID":"sha256:0f43c31aa3359159d4557dad3cfaf812d8ce44db9cb9ae970e06d3479070b660","firstCommand":"VvlllczXFX2z","firstArg":"47DEQpj8HBSa"},{"imageID":"sha256:0f43c31aa3359159d4557dad3cfaf812d8ce44db9cb9ae970e06d3479070b660","firstCommand":"VvlllczXFX2z","firstArg":"Yfg8wToYnj2U"}]}]},"QhwI_DkC7RkW":{"count":2,"shapes":[{"restartAlways":true,"containers":[{"imageID":"sha256:c69ffd8f8dcceedc2d6eb306cea33f8beabc1be1308cd5f4ee8b9a8e3eab9843","firstCommand":"N9KxLV2avCo2"},{"imageID":"sha256:e3e65409fc2b27ad0aaeb500a39e264663d2980821f099b830b551785ce4ce8b","firstCommand":"N9KxLV2avCo2"}]},{"restartAlways":true,"initContainers":[{"imageID":"sha256:e3e65409fc2b27ad0aaeb500a39e264663d2980821f099b830b551785ce4ce8b","firstCommand":"N9KxLV2avCo2"}],"containers":[{"imageID":"sha256:c69ffd8f8dcceedc2d6eb306cea33f8beabc1be1308cd5f4ee8b9a8e3eab9843","firstCommand":"N9KxLV2avCo2"},{"imageID":"sha256:c69ffd8f8dcceedc2d6eb306cea33f8beabc1be1308cd5f4ee8b9a8e3eab9843","firstCommand":"N9KxLV2avCo2"},{"imageID":"sha256:e3e65409fc2b27ad0aaeb500a39e264663d2980821f099b830b551785ce4ce8b","firstCommand":"N9KxLV2avCo2"},{"imageID":"sha256:e3e65409fc2b27ad0aaeb500a39e264663d2980821f099b830b551785ce4ce8b","firstCommand":"N9KxLV2avCo2"},{"imageID":"sha256:e3e65409fc2b27ad0aaeb500a39e264663d2980821f099b830b551785ce4ce8b","firstCommand":"N9KxLV2avCo2"},{"imageID":"sha256:e3e65409fc2b27ad0aaeb500a39e264663d2980821f099b830b551785ce4ce8b","firstCommand":"N9KxLV2avCo2"},{"imageID":"sha256:e3e65409fc2b27ad0aaeb500a39e264663d2980821f099b830b551785ce4ce8b","firstCommand":"N9KxLV2avCo2"},{"imageID":"sha256:e3e65409fc2b27ad0aaeb500a39e264663d2980821f099b830b551785ce4ce8b","firstCommand":"N9KxLV2avCo2"}]}]},"RGIC92MKxXzi":{"count":1,"shapes":[{"restartAlways":true,"containers":[{"imageID":"sha256:c69ffd8f8dcceedc2d6eb306cea33f8beabc1be1308cd5f4ee8b9a8e3eab9843","firstArg":"aV86_fLS8Aye"},{"imageID":"sha256:8cc27777e72233024fe84ee1faa168aec715a0b24912a3ce70715ddccba328df","firstCommand":"RxKEaTc90uWM","firstArg":"r7zEsGNvlNrk"}]}]},"WPOX0dN4tqDv":{"count":1,"shapes":[{"restartAlways":true,"containers":[{"imageID":"sha256:831f30660844091d6154e2674d3a9da6f34271bf8a2c40b56f7416066318742b","firstCommand":"9RqZYN0XQJuD"},{"imageID":"sha256:c69ffd8f8dcceedc2d6eb306cea33f8beabc1be1308cd5f4ee8b9a8e3eab9843","firstArg":"47DEQpj8HBSa"}]}]},"XStfgPOmHYzB":{"count":1,"shapes":[{"restartAlways":true,"containers":[{"imageID":"sha256:75d996f6147edb88c09fd1a052099de66638590d7d03a735006244bc9e19f898","firstCommand":"AiycEq1CClLy","firstArg":"r7zEsGNvlNrk"}]}]},"YgQZCjyI6D9c":{"count":1,"terminalCount":1,"shapes":[]},"cMIdS73ad1cT":{"count":1,"shapes":[{"restartAlways":true,"containers":[{"imageID":"sha256:f4d4282cb53325e737ad68abbfcb70687ae04fb50353f4f0ba0ba5703b15009a","firstCommand":"09D_XsQaIohy","firstArg":"r7zEsGNvlNrk"}]}]},"cWNZj9dm5f4i":{"count":1,"shapes":[{"restartAlways":true,"containers":[{"imageID":"sha256:f042fa25014f3d37f3ea967d21f361d2a11833ae18f2c750318101b25d2497ce","firstCommand":"WFJEBSInAJpw","firstArg":"r7zEsGNvlNrk"}]}]},"coRkJ8lvmtSk":{"count":1,"ignoredCount":1,"shapes":[]},"de1zmznD1F_L":{"count":1,"shapes":[{"restartAlways":true,"containers":[{"imageID":"sha256:8eabac819f289e29d75c7ab172d8124554849a47f0b00770928c3eb19a5a31c4","firstCommand":"rb3iLnl9GmZR","firstArg":"r7zEsGNvlNrk"}]}]},"eCjF12pGOLUF":{"count":1,"shapes":[{"restartAlways":true,"containers":[{"imageID":"sha256:dd38b8be3af889b0f97e2df41517c89a11260901432a9a1ee943195bb3a22737","firstCommand":"N9KxLV2avCo2"},{"imageID":"sha256:dd38b8be3af889b0f97e2df41517c89a11260901432a9a1ee943195bb3a22737","firstCommand":"N9KxLV2avCo2"},{"imageID":"sha256:c69ffd8f8dcceedc2d6eb306cea33f8beabc1be1308cd5f4ee8b9a8e3eab9843","firstArg":"aV86_fLS8Aye"}]}]},"ecUe5cl6bPwX":{"count":4,"shapes":[{"restartAlways":true,"containers":[{"imageID":"sha256:72bbe2c638872937108f647950ab8ad35c0428ca8ecc6a39a8314aace7d95078","firstCommand":"pgaYl1gqTV2L","firstArg":"47DEQpj8HBSa"},{"imageID":"sha256:c69ffd8f8dcceedc2d6eb306cea33f8beabc1be1308cd5f4ee8b9a8e3eab9843","firstArg":"aV86_fLS8Aye"}]},{"restartAlways":true,"containers":[{"imageID":"sha256:c69ffd8f8dcceedc2d6eb306cea33f8beabc1be1308cd5f4ee8b9a8e3eab9843","firstArg":"aV86_fLS8Aye"},{"imageID":"sha256:a92c310ce30dcb3de85d6aac868e0d80919670fa29ef83d55edd96b0cae35563","firstCommand":"kwBOP3_4oVMx","firstArg":"47DEQpj8HBSa"}]},{"restartAlways":true,"containers":[{"imageID":"sha256:fd3e9f8f00a59bda7483ec7dc8a0ed602f9ca30e3d72b22072dbdf2819da3f61","firstCommand":"buSkac1OkQU4","firstArg":"gcNszUTvGLqr"}]},{"restartAlways":true,"containers":[{"imageID":"sha256:c69ffd8f8dcceedc2d6eb306cea33f8beabc1be1308cd5f4ee8b9a8e3eab9843","firstArg":"aV86_fLS8Aye"},{"imageID":"sha256:7c2431a990bcddde98829abda81950247021a2ebbabc964b1516ea046b5f1d4e","firstCommand":"Tt12uskXp-TN","firstArg":"zO0oxtw_mcI5"}]}]},"inQbBSWydCWk":{"count":1,"shapes":[{"restartAlways":true,"containers":[{"imageID":"sha256:e85850a4ae1a1e3ec2c590a4936d640882b6550124da22031c85b526afbf52df","firstCommand":"N9KxLV2avCo2","firstArg":"k19oMZ1PIn4C"}]}]},"ioFx0itCFPcs":{"count":1,"shapes":[{"restartAlways":true,"initContainers":[{"imageID":"sha256:3f0aa9cd04713acc5c6fea721bd849e1500da8ae945e0b32000887f34d786e0b","firstCommand":"icTsn2s_EIax","firstArg":"47DEQpj8HBSa"},{"imageID":"sha256:1f952cec1e5332b84bdffa249cd426f39087058d6544ddcec650a414c15a9b68","firstCommand":"icTsn2s_EIax","firstArg":"47DEQpj8HBSa"}],"containers":[{"imageID":"sha256:86af77350cfe6fd69280157e4162aa0147873d9431c641ae4ad3e881ff768a73","firstCommand":"bFMn8sKmxXTo","firstArg":"zO0oxtw_mcI5"}]}]},"jmuesrd3LOV1":{"count":2,"shapes":[{"restartAlways":true,"initContainers":[{"imageID":"sha256:c69ffd8f8dcceedc2d6eb306cea33f8beabc1be1308cd5f4ee8b9a8e3eab9843","firstCommand":"N9KxLV2avCo2","firstArg":"CSx56PgOVZ5A"}],"containers":[{"imageID":"sha256:c69ffd8f8dcceedc2d6eb306cea33f8beabc1be1308cd5f4ee8b9a8e3eab9843","firstArg":"aV86_fLS8Aye"}]},{"restartAlways":true,"containers":[{"imageID":"sha256:c69ffd8f8dcceedc2d6eb306cea33f8beabc1be1308cd5f4ee8b9a8e3eab9843","firstArg":"aV86_fLS8Aye"},{"imageID":"sha256:b472823604757237c2d16bd6f6221f4cf562aa3b05942c7f602e1e8b2e55a7c6","firstArg":"zO0oxtw_mcI5"}]}]},"k2lPYOlQaXUX":{"count":1,"shapes":[{"restartAlways":true,"containers":[{"imageID":"sha256:8375671da86aa527ee7e291d86971b0baa823ffc7663b5a983084456e76c0f59","firstCommand":"XqE_QmN1P920","firstArg":"r7zEsGNvlNrk"}]}]},"n8ih8iFf25K7":{"count":2,"shapes":[{"restartAlways":true,"containers":[{"imageID":"sha256:79f99fd6cce984287932edf0d009660bb488d663081f3d62ec3b23bc8bfbf6c2","firstCommand":"CX8KiHZjJyVD"}]},{"restartAlways":true,"containers":[{"imageID":"sha256:9724d2036305cbd729e1f484c5bad89971de977fff8a6723fef1873858dd1123","firstCommand":"N9KxLV2avCo2"}]}]},"nMMaB9HQd060":{"count":2,"ignoredCount":1,"shapes":[{"restartAlways":true,"containers":[{"imageID":"sha256:c3a77aa4d03b89ea284e3467a268e5989a77a2ef63e685eb1d5c5ea5b3922b7a","firstArg":"aZFPdeh8u0QY"}]}]},"nO8YkKLJMsaW":{"count":1,"shapes":[{"restartAlways":true,"initContainers":[{"imageID":"sha256:df606f3b71d4376d1a2108c09f0d3dab455fc30bcb67c60e91590c105e9025bf","firstCommand":"icTsn2s_EIax"}],"containers":[{"imageID":"sha256:df606f3b71d4376d1a2108c09f0d3dab455fc30bcb67c60e91590c105e9025bf","firstCommand":"N9KxLV2avCo2","firstArg":"k19oMZ1PIn4C"},{"imageID":"sha256:0f43c31aa3359159d4557dad3cfaf812d8ce44db9cb9ae970e06d3479070b660","firstCommand":"VvlllczXFX2z","firstArg":"47DEQpj8HBSa"}]}]},"nTf6g7hF1YMN":{"count":1,"shapes":[{"restartAlways":true,"containers":[{"imageID":"sha256:aa24edce3d740f84c40018e94cdbf2bc7375268d13d57c2d664e43a46ccea3fc","firstArg":"8DZ1gO9Xqn6d"}]}]},"oNeg9VIL3k3N":{"count":1,"shapes":[{"restartAlways":true,"containers":[{"imageID":"sha256:eddedae7578d79b5a3f748000ae5c00b9f14a04710f9f9ec7b52fc569be5dfb8","firstCommand":"oNeg9VIL3k3N","firstArg":"r7zEsGNvlNrk"}]}]},"pepxDbsB4SPa":{"count":1,"shapes":[{"restartAlways":true,"containers":[{"imageID":"sha256:0f43c31aa3359159d4557dad3cfaf812d8ce44db9cb9ae970e06d3479070b660","firstCommand":"VvlllczXFX2z","firstArg":"r7zEsGNvlNrk"}]}]},"q1TpN3EUiKjG":{"count":1,"shapes":[{"restartAlways":true,"containers":[{"imageID":"sha256:a824e468cf8dd61d347e35b2ee5bc2f815666957647098e21a1bb56ff613e5b9","firstCommand":"FATPL2W-sitV","firstArg":"r7zEsGNvlNrk"}]}]},"vDGHaQt8x6K_":{"count":1,"shapes":[{"restartAlways":true,"containers":[{"imageID":"sha256:8c885ea0b3c5124989f0a9b93eba98eb9fca6bbd0262772d85d90bf713a4d572","firstArg":"zO0oxtw_mcI5"}]}]}}}packageserver_current.log0000640000000000000000000010776715114274105033755 0ustar0000000000000000config/pod/openshift-operator-lifecycle-manager/logs/packageserver-675f5c767c-mtdrq2025-12-04T11:45:16.007472051Z I1204 11:45:16.007457 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController 2025-12-04T11:45:16.007479361Z I1204 11:45:16.007468 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file 2025-12-04T11:45:16.008117693Z I1204 11:45:16.008085 1 secure_serving.go:213] Serving securely on [::]:5443 2025-12-04T11:45:16.008163424Z I1204 11:45:16.008139 1 tlsconfig.go:243] "Starting DynamicServingCertificateController" 2025-12-04T11:45:16.008210396Z I1204 11:45:16.008132 1 dynamic_serving_content.go:135] "Starting controller" name="serving-cert::apiserver.local.config/certificates/apiserver.crt::apiserver.local.config/certificates/apiserver.key" 2025-12-04T11:45:16.107958813Z I1204 11:45:16.107873 1 shared_informer.go:320] Caches are synced for client-ca::kube-system::extension-apiserver-authentication::client-ca-file 2025-12-04T11:45:16.107958813Z I1204 11:45:16.107907 1 shared_informer.go:320] Caches are synced for RequestHeaderAuthRequestController 2025-12-04T11:45:16.107996574Z I1204 11:45:16.107912 1 shared_informer.go:320] Caches are synced for RequestHeaderAuthRequestController 2025-12-04T11:45:16.108083457Z I1204 11:45:16.108036 1 shared_informer.go:320] Caches are synced for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file 2025-12-04T11:45:16.108091617Z I1204 11:45:16.108070 1 shared_informer.go:320] Caches are synced for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file 2025-12-04T11:45:16.108099308Z I1204 11:45:16.108089 1 shared_informer.go:320] Caches are synced for client-ca::kube-system::extension-apiserver-authentication::client-ca-file 2025-12-04T11:45:17.007819023Z W1204 11:45:17.007737 1 logging.go:55] [core] [Channel #4 SubChannel #8]grpc: addrConn.createTransport failed to connect to {Addr: "redhat-marketplace.openshift-marketplace.svc:50051", ServerName: "redhat-marketplace.openshift-marketplace.svc:50051", }. Err: connection error: desc = "transport: Error while dialing: dial tcp 172.30.165.235:50051: connect: connection refused" 2025-12-04T11:45:17.007921936Z W1204 11:45:17.007873 1 logging.go:55] [core] [Channel #1 SubChannel #3]grpc: addrConn.createTransport failed to connect to {Addr: "certified-operators.openshift-marketplace.svc:50051", ServerName: "certified-operators.openshift-marketplace.svc:50051", }. Err: connection error: desc = "transport: Error while dialing: dial tcp 172.30.138.0:50051: connect: connection refused" 2025-12-04T11:45:17.007967058Z W1204 11:45:17.007892 1 logging.go:55] [core] [Channel #2 SubChannel #5]grpc: addrConn.createTransport failed to connect to {Addr: "community-operators.openshift-marketplace.svc:50051", ServerName: "community-operators.openshift-marketplace.svc:50051", }. Err: connection error: desc = "transport: Error while dialing: dial tcp 172.30.111.58:50051: connect: connection refused" 2025-12-04T11:45:17.008263548Z W1204 11:45:17.008227 1 logging.go:55] [core] [Channel #6 SubChannel #7]grpc: addrConn.createTransport failed to connect to {Addr: "redhat-operators.openshift-marketplace.svc:50051", ServerName: "redhat-operators.openshift-marketplace.svc:50051", }. Err: connection error: desc = "transport: Error while dialing: dial tcp 172.30.86.197:50051: connect: connection refused" 2025-12-04T11:45:18.620432565Z W1204 11:45:18.620315 1 logging.go:55] [core] [Channel #1 SubChannel #3]grpc: addrConn.createTransport failed to connect to {Addr: "certified-operators.openshift-marketplace.svc:50051", ServerName: "certified-operators.openshift-marketplace.svc:50051", }. Err: connection error: desc = "transport: Error while dialing: dial tcp 172.30.138.0:50051: connect: connection refused" 2025-12-04T11:45:18.738708137Z W1204 11:45:18.738632 1 logging.go:55] [core] [Channel #4 SubChannel #8]grpc: addrConn.createTransport failed to connect to {Addr: "redhat-marketplace.openshift-marketplace.svc:50051", ServerName: "redhat-marketplace.openshift-marketplace.svc:50051", }. Err: connection error: desc = "transport: Error while dialing: dial tcp 172.30.165.235:50051: connect: connection refused" 2025-12-04T11:45:18.846669561Z W1204 11:45:18.846593 1 logging.go:55] [core] [Channel #2 SubChannel #5]grpc: addrConn.createTransport failed to connect to {Addr: "community-operators.openshift-marketplace.svc:50051", ServerName: "community-operators.openshift-marketplace.svc:50051", }. Err: connection error: desc = "transport: Error while dialing: dial tcp 172.30.111.58:50051: connect: connection refused" 2025-12-04T11:45:18.886229623Z W1204 11:45:18.886151 1 logging.go:55] [core] [Channel #6 SubChannel #7]grpc: addrConn.createTransport failed to connect to {Addr: "redhat-operators.openshift-marketplace.svc:50051", ServerName: "redhat-operators.openshift-marketplace.svc:50051", }. Err: connection error: desc = "transport: Error while dialing: dial tcp 172.30.86.197:50051: connect: connection refused" 2025-12-04T11:45:20.811616937Z W1204 11:45:20.811528 1 logging.go:55] [core] [Channel #4 SubChannel #8]grpc: addrConn.createTransport failed to connect to {Addr: "redhat-marketplace.openshift-marketplace.svc:50051", ServerName: "redhat-marketplace.openshift-marketplace.svc:50051", }. Err: connection error: desc = "transport: Error while dialing: dial tcp 172.30.165.235:50051: connect: connection refused" 2025-12-04T11:45:20.990878530Z W1204 11:45:20.990791 1 logging.go:55] [core] [Channel #1 SubChannel #3]grpc: addrConn.createTransport failed to connect to {Addr: "certified-operators.openshift-marketplace.svc:50051", ServerName: "certified-operators.openshift-marketplace.svc:50051", }. Err: connection error: desc = "transport: Error while dialing: dial tcp 172.30.138.0:50051: connect: connection refused" 2025-12-04T11:45:21.163663032Z W1204 11:45:21.163548 1 logging.go:55] [core] [Channel #2 SubChannel #5]grpc: addrConn.createTransport failed to connect to {Addr: "community-operators.openshift-marketplace.svc:50051", ServerName: "community-operators.openshift-marketplace.svc:50051", }. Err: connection error: desc = "transport: Error while dialing: dial tcp 172.30.111.58:50051: connect: connection refused" 2025-12-04T11:45:21.929504346Z W1204 11:45:21.929393 1 logging.go:55] [core] [Channel #6 SubChannel #7]grpc: addrConn.createTransport failed to connect to {Addr: "redhat-operators.openshift-marketplace.svc:50051", ServerName: "redhat-operators.openshift-marketplace.svc:50051", }. Err: connection error: desc = "transport: Error while dialing: dial tcp 172.30.86.197:50051: connect: connection refused" 2025-12-04T11:45:22.794226759Z time="2025-12-04T11:45:22Z" level=info msg="updating PackageManifest based on CatalogSource changes: {community-operators openshift-marketplace}" action="sync catalogsource" address="community-operators.openshift-marketplace.svc:50051" name=community-operators namespace=openshift-marketplace 2025-12-04T11:45:22.794311822Z time="2025-12-04T11:45:22Z" level=warning msg="error getting bundle stream" action="refresh cache" err="rpc error: code = Unavailable desc = connection error: desc = \"transport: Error while dialing: dial tcp 172.30.111.58:50051: connect: connection refused\"" source="{community-operators openshift-marketplace}" 2025-12-04T11:45:23.214454820Z time="2025-12-04T11:45:23Z" level=info msg="updating PackageManifest based on CatalogSource changes: {redhat-marketplace openshift-marketplace}" action="sync catalogsource" address="redhat-marketplace.openshift-marketplace.svc:50051" name=redhat-marketplace namespace=openshift-marketplace 2025-12-04T11:45:23.214454820Z time="2025-12-04T11:45:23Z" level=warning msg="error getting bundle stream" action="refresh cache" err="rpc error: code = Unavailable desc = connection error: desc = \"transport: Error while dialing: dial tcp 172.30.165.235:50051: connect: connection refused\"" source="{redhat-marketplace openshift-marketplace}" 2025-12-04T11:45:24.960496248Z W1204 11:45:24.960398 1 logging.go:55] [core] [Channel #2 SubChannel #5]grpc: addrConn.createTransport failed to connect to {Addr: "community-operators.openshift-marketplace.svc:50051", ServerName: "community-operators.openshift-marketplace.svc:50051", }. Err: connection error: desc = "transport: Error while dialing: dial tcp 172.30.111.58:50051: connect: connection refused" 2025-12-04T11:45:25.588685818Z W1204 11:45:25.588593 1 logging.go:55] [core] [Channel #4 SubChannel #8]grpc: addrConn.createTransport failed to connect to {Addr: "redhat-marketplace.openshift-marketplace.svc:50051", ServerName: "redhat-marketplace.openshift-marketplace.svc:50051", }. Err: connection error: desc = "transport: Error while dialing: dial tcp 172.30.165.235:50051: connect: connection refused" 2025-12-04T11:45:25.881383065Z W1204 11:45:25.881291 1 logging.go:55] [core] [Channel #1 SubChannel #3]grpc: addrConn.createTransport failed to connect to {Addr: "certified-operators.openshift-marketplace.svc:50051", ServerName: "certified-operators.openshift-marketplace.svc:50051", }. Err: connection error: desc = "transport: Error while dialing: dial tcp 172.30.138.0:50051: connect: connection refused" 2025-12-04T11:45:26.417774348Z W1204 11:45:26.417671 1 logging.go:55] [core] [Channel #6 SubChannel #7]grpc: addrConn.createTransport failed to connect to {Addr: "redhat-operators.openshift-marketplace.svc:50051", ServerName: "redhat-operators.openshift-marketplace.svc:50051", }. Err: connection error: desc = "transport: Error while dialing: dial tcp 172.30.86.197:50051: connect: connection refused" 2025-12-04T11:45:30.324750098Z W1204 11:45:30.324659 1 logging.go:55] [core] [Channel #2 SubChannel #5]grpc: addrConn.createTransport failed to connect to {Addr: "community-operators.openshift-marketplace.svc:50051", ServerName: "community-operators.openshift-marketplace.svc:50051", }. Err: connection error: desc = "transport: Error while dialing: dial tcp 172.30.111.58:50051: connect: connection refused" 2025-12-04T11:45:30.441449609Z time="2025-12-04T11:45:30Z" level=info msg="updating PackageManifest based on CatalogSource changes: {redhat-marketplace openshift-marketplace}" action="sync catalogsource" address="redhat-marketplace.openshift-marketplace.svc:50051" name=redhat-marketplace namespace=openshift-marketplace 2025-12-04T11:45:30.441449609Z time="2025-12-04T11:45:30Z" level=warning msg="error getting bundle stream" action="refresh cache" err="rpc error: code = Unavailable desc = connection error: desc = \"transport: Error while dialing: dial tcp 172.30.165.235:50051: connect: connection refused\"" source="{redhat-marketplace openshift-marketplace}" 2025-12-04T11:45:32.326500875Z W1204 11:45:32.326369 1 logging.go:55] [core] [Channel #4 SubChannel #8]grpc: addrConn.createTransport failed to connect to {Addr: "redhat-marketplace.openshift-marketplace.svc:50051", ServerName: "redhat-marketplace.openshift-marketplace.svc:50051", }. Err: connection error: desc = "transport: Error while dialing: dial tcp 172.30.165.235:50051: connect: connection refused" 2025-12-04T11:45:33.406263550Z W1204 11:45:33.406189 1 logging.go:55] [core] [Channel #6 SubChannel #7]grpc: addrConn.createTransport failed to connect to {Addr: "redhat-operators.openshift-marketplace.svc:50051", ServerName: "redhat-operators.openshift-marketplace.svc:50051", }. Err: connection error: desc = "transport: Error while dialing: dial tcp 172.30.86.197:50051: connect: connection refused" 2025-12-04T11:45:33.560628060Z W1204 11:45:33.558735 1 logging.go:55] [core] [Channel #1 SubChannel #3]grpc: addrConn.createTransport failed to connect to {Addr: "certified-operators.openshift-marketplace.svc:50051", ServerName: "certified-operators.openshift-marketplace.svc:50051", }. Err: connection error: desc = "transport: Error while dialing: dial tcp 172.30.138.0:50051: connect: connection refused" 2025-12-04T11:45:41.179625669Z W1204 11:45:41.179554 1 logging.go:55] [core] [Channel #2 SubChannel #5]grpc: addrConn.createTransport failed to connect to {Addr: "community-operators.openshift-marketplace.svc:50051", ServerName: "community-operators.openshift-marketplace.svc:50051", }. Err: connection error: desc = "transport: Error while dialing: dial tcp 172.30.111.58:50051: connect: connection refused" 2025-12-04T11:45:43.977378806Z W1204 11:45:43.977304 1 logging.go:55] [core] [Channel #6 SubChannel #7]grpc: addrConn.createTransport failed to connect to {Addr: "redhat-operators.openshift-marketplace.svc:50051", ServerName: "redhat-operators.openshift-marketplace.svc:50051", }. Err: connection error: desc = "transport: Error while dialing: dial tcp 172.30.86.197:50051: connect: connection refused" 2025-12-04T11:45:44.445077346Z W1204 11:45:44.445026 1 logging.go:55] [core] [Channel #4 SubChannel #8]grpc: addrConn.createTransport failed to connect to {Addr: "redhat-marketplace.openshift-marketplace.svc:50051", ServerName: "redhat-marketplace.openshift-marketplace.svc:50051", }. Err: connection error: desc = "transport: Error while dialing: dial tcp 172.30.165.235:50051: connect: connection refused" 2025-12-04T11:45:45.949470224Z W1204 11:45:45.949388 1 logging.go:55] [core] [Channel #1 SubChannel #3]grpc: addrConn.createTransport failed to connect to {Addr: "certified-operators.openshift-marketplace.svc:50051", ServerName: "certified-operators.openshift-marketplace.svc:50051", }. Err: connection error: desc = "transport: Error while dialing: dial tcp 172.30.138.0:50051: connect: connection refused" 2025-12-04T11:45:55.036576927Z W1204 11:45:55.033301 1 logging.go:55] [core] [Channel #2 SubChannel #5]grpc: addrConn.createTransport failed to connect to {Addr: "community-operators.openshift-marketplace.svc:50051", ServerName: "community-operators.openshift-marketplace.svc:50051", }. Err: connection error: desc = "transport: Error while dialing: dial tcp 172.30.111.58:50051: connect: connection refused" 2025-12-04T11:45:59.942198975Z W1204 11:45:59.942105 1 logging.go:55] [core] [Channel #6 SubChannel #7]grpc: addrConn.createTransport failed to connect to {Addr: "redhat-operators.openshift-marketplace.svc:50051", ServerName: "redhat-operators.openshift-marketplace.svc:50051", }. Err: connection error: desc = "transport: Error while dialing: dial tcp 172.30.86.197:50051: connect: connection refused" 2025-12-04T11:46:02.518216373Z W1204 11:46:02.518097 1 logging.go:55] [core] [Channel #4 SubChannel #8]grpc: addrConn.createTransport failed to connect to {Addr: "redhat-marketplace.openshift-marketplace.svc:50051", ServerName: "redhat-marketplace.openshift-marketplace.svc:50051", }. Err: connection error: desc = "transport: Error while dialing: dial tcp 172.30.165.235:50051: connect: connection refused" 2025-12-04T11:46:05.265682947Z W1204 11:46:05.265610 1 logging.go:55] [core] [Channel #1 SubChannel #3]grpc: addrConn.createTransport failed to connect to {Addr: "certified-operators.openshift-marketplace.svc:50051", ServerName: "certified-operators.openshift-marketplace.svc:50051", }. Err: connection error: desc = "transport: Error while dialing: dial tcp 172.30.138.0:50051: connect: connection refused" 2025-12-04T11:46:17.040198002Z W1204 11:46:17.040128 1 logging.go:55] [core] [Channel #2 SubChannel #5]grpc: addrConn.createTransport failed to connect to {Addr: "community-operators.openshift-marketplace.svc:50051", ServerName: "community-operators.openshift-marketplace.svc:50051", }. Err: connection error: desc = "transport: Error while dialing: dial tcp 172.30.111.58:50051: connect: connection refused" 2025-12-04T11:46:21.562183341Z W1204 11:46:21.560968 1 logging.go:55] [core] [Channel #6 SubChannel #7]grpc: addrConn.createTransport failed to connect to {Addr: "redhat-operators.openshift-marketplace.svc:50051", ServerName: "redhat-operators.openshift-marketplace.svc:50051", }. Err: connection error: desc = "transport: Error while dialing: dial tcp 172.30.86.197:50051: connect: connection refused" 2025-12-04T11:46:26.625627950Z W1204 11:46:26.625485 1 logging.go:55] [core] [Channel #4 SubChannel #8]grpc: addrConn.createTransport failed to connect to {Addr: "redhat-marketplace.openshift-marketplace.svc:50051", ServerName: "redhat-marketplace.openshift-marketplace.svc:50051", }. Err: connection error: desc = "transport: Error while dialing: dial tcp 172.30.165.235:50051: connect: connection refused" 2025-12-04T11:46:29.163333870Z W1204 11:46:29.163237 1 logging.go:55] [core] [Channel #1 SubChannel #3]grpc: addrConn.createTransport failed to connect to {Addr: "certified-operators.openshift-marketplace.svc:50051", ServerName: "certified-operators.openshift-marketplace.svc:50051", }. Err: connection error: desc = "transport: Error while dialing: dial tcp 172.30.138.0:50051: connect: connection refused" 2025-12-04T11:46:56.438910511Z W1204 11:46:56.438828 1 logging.go:55] [core] [Channel #6 SubChannel #7]grpc: addrConn.createTransport failed to connect to {Addr: "redhat-operators.openshift-marketplace.svc:50051", ServerName: "redhat-operators.openshift-marketplace.svc:50051", }. Err: connection error: desc = "transport: Error while dialing: dial tcp 172.30.86.197:50051: connect: connection refused" 2025-12-04T11:47:04.677377043Z W1204 11:47:04.677285 1 logging.go:55] [core] [Channel #1 SubChannel #3]grpc: addrConn.createTransport failed to connect to {Addr: "certified-operators.openshift-marketplace.svc:50051", ServerName: "certified-operators.openshift-marketplace.svc:50051", }. Err: connection error: desc = "transport: Error while dialing: dial tcp 172.30.138.0:50051: connect: connection refused" 2025-12-04T11:47:04.884564115Z W1204 11:47:04.884473 1 logging.go:55] [core] [Channel #2 SubChannel #5]grpc: addrConn.createTransport failed to connect to {Addr: "community-operators.openshift-marketplace.svc:50051", ServerName: "community-operators.openshift-marketplace.svc:50051", }. Err: connection error: desc = "transport: Error while dialing: dial tcp 172.30.111.58:50051: connect: connection refused" 2025-12-04T11:47:10.894957843Z W1204 11:47:10.894854 1 logging.go:55] [core] [Channel #4 SubChannel #8]grpc: addrConn.createTransport failed to connect to {Addr: "redhat-marketplace.openshift-marketplace.svc:50051", ServerName: "redhat-marketplace.openshift-marketplace.svc:50051", }. Err: connection error: desc = "transport: Error while dialing: dial tcp 172.30.165.235:50051: connect: connection refused" 2025-12-04T11:48:00.902192459Z W1204 11:48:00.902077 1 logging.go:55] [core] [Channel #6 SubChannel #7]grpc: addrConn.createTransport failed to connect to {Addr: "redhat-operators.openshift-marketplace.svc:50051", ServerName: "redhat-operators.openshift-marketplace.svc:50051", }. Err: connection error: desc = "transport: Error while dialing: dial tcp 172.30.86.197:50051: connect: connection refused" 2025-12-04T11:48:11.719199056Z W1204 11:48:11.719114 1 logging.go:55] [core] [Channel #1 SubChannel #3]grpc: addrConn.createTransport failed to connect to {Addr: "certified-operators.openshift-marketplace.svc:50051", ServerName: "certified-operators.openshift-marketplace.svc:50051", }. Err: connection error: desc = "transport: Error while dialing: dial tcp 172.30.138.0:50051: connect: connection refused" 2025-12-04T11:48:13.369580242Z W1204 11:48:13.369447 1 logging.go:55] [core] [Channel #2 SubChannel #5]grpc: addrConn.createTransport failed to connect to {Addr: "community-operators.openshift-marketplace.svc:50051", ServerName: "community-operators.openshift-marketplace.svc:50051", }. Err: connection error: desc = "transport: Error while dialing: dial tcp 172.30.111.58:50051: connect: connection refused" 2025-12-04T11:48:14.021382897Z W1204 11:48:14.021254 1 logging.go:55] [core] [Channel #4 SubChannel #8]grpc: addrConn.createTransport failed to connect to {Addr: "redhat-marketplace.openshift-marketplace.svc:50051", ServerName: "redhat-marketplace.openshift-marketplace.svc:50051", }. Err: connection error: desc = "transport: Error while dialing: dial tcp 172.30.165.235:50051: connect: connection refused" 2025-12-04T11:49:59.231054154Z W1204 11:49:59.230867 1 logging.go:55] [core] [Channel #6 SubChannel #7]grpc: addrConn.createTransport failed to connect to {Addr: "redhat-operators.openshift-marketplace.svc:50051", ServerName: "redhat-operators.openshift-marketplace.svc:50051", }. Err: connection error: desc = "transport: Error while dialing: dial tcp 172.30.86.197:50051: connect: connection refused" 2025-12-04T11:50:00.103060210Z W1204 11:50:00.102974 1 logging.go:55] [core] [Channel #2 SubChannel #5]grpc: addrConn.createTransport failed to connect to {Addr: "community-operators.openshift-marketplace.svc:50051", ServerName: "community-operators.openshift-marketplace.svc:50051", }. Err: connection error: desc = "transport: Error while dialing: dial tcp 172.30.111.58:50051: connect: connection refused" 2025-12-04T11:50:19.227734493Z W1204 11:50:19.227077 1 logging.go:55] [core] [Channel #1 SubChannel #3]grpc: addrConn.createTransport failed to connect to {Addr: "certified-operators.openshift-marketplace.svc:50051", ServerName: "certified-operators.openshift-marketplace.svc:50051", }. Err: connection error: desc = "transport: Error while dialing: dial tcp 172.30.138.0:50051: connect: connection refused" 2025-12-04T11:50:22.608415154Z W1204 11:50:22.608320 1 logging.go:55] [core] [Channel #4 SubChannel #8]grpc: addrConn.createTransport failed to connect to {Addr: "redhat-marketplace.openshift-marketplace.svc:50051", ServerName: "redhat-marketplace.openshift-marketplace.svc:50051", }. Err: connection error: desc = "transport: Error while dialing: dial tcp 172.30.165.235:50051: connect: connection refused" 2025-12-04T11:51:43.361077118Z W1204 11:51:43.360952 1 logging.go:55] [core] [Channel #2 SubChannel #5]grpc: addrConn.createTransport failed to connect to {Addr: "community-operators.openshift-marketplace.svc:50051", ServerName: "community-operators.openshift-marketplace.svc:50051", }. Err: connection error: desc = "transport: Error while dialing: dial tcp 172.30.111.58:50051: connect: connection refused" 2025-12-04T11:52:08.922232479Z W1204 11:52:08.922141 1 logging.go:55] [core] [Channel #6 SubChannel #7]grpc: addrConn.createTransport failed to connect to {Addr: "redhat-operators.openshift-marketplace.svc:50051", ServerName: "redhat-operators.openshift-marketplace.svc:50051", }. Err: connection error: desc = "transport: Error while dialing: dial tcp 172.30.86.197:50051: connect: connection refused" 2025-12-04T11:52:10.920150418Z W1204 11:52:10.920022 1 logging.go:55] [core] [Channel #1 SubChannel #3]grpc: addrConn.createTransport failed to connect to {Addr: "certified-operators.openshift-marketplace.svc:50051", ServerName: "certified-operators.openshift-marketplace.svc:50051", }. Err: connection error: desc = "transport: Error while dialing: dial tcp 172.30.138.0:50051: connect: connection refused" 2025-12-04T11:52:27.264264795Z W1204 11:52:27.264123 1 logging.go:55] [core] [Channel #4 SubChannel #8]grpc: addrConn.createTransport failed to connect to {Addr: "redhat-marketplace.openshift-marketplace.svc:50051", ServerName: "redhat-marketplace.openshift-marketplace.svc:50051", }. Err: connection error: desc = "transport: Error while dialing: dial tcp 172.30.165.235:50051: connect: connection refused" 2025-12-04T11:53:56.744757168Z W1204 11:53:56.744662 1 logging.go:55] [core] [Channel #2 SubChannel #5]grpc: addrConn.createTransport failed to connect to {Addr: "community-operators.openshift-marketplace.svc:50051", ServerName: "community-operators.openshift-marketplace.svc:50051", }. Err: connection error: desc = "transport: Error while dialing: dial tcp 172.30.111.58:50051: connect: connection refused" 2025-12-04T11:54:05.787046875Z W1204 11:54:05.786959 1 logging.go:55] [core] [Channel #4 SubChannel #8]grpc: addrConn.createTransport failed to connect to {Addr: "redhat-marketplace.openshift-marketplace.svc:50051", ServerName: "redhat-marketplace.openshift-marketplace.svc:50051", }. Err: connection error: desc = "transport: Error while dialing: dial tcp 172.30.165.235:50051: connect: connection refused" 2025-12-04T11:54:07.159984675Z W1204 11:54:07.159873 1 logging.go:55] [core] [Channel #1 SubChannel #3]grpc: addrConn.createTransport failed to connect to {Addr: "certified-operators.openshift-marketplace.svc:50051", ServerName: "certified-operators.openshift-marketplace.svc:50051", }. Err: connection error: desc = "transport: Error while dialing: dial tcp 172.30.138.0:50051: connect: connection refused" 2025-12-04T11:54:30.995995403Z W1204 11:54:30.995877 1 logging.go:55] [core] [Channel #6 SubChannel #7]grpc: addrConn.createTransport failed to connect to {Addr: "redhat-operators.openshift-marketplace.svc:50051", ServerName: "redhat-operators.openshift-marketplace.svc:50051", }. Err: connection error: desc = "transport: Error while dialing: dial tcp 172.30.86.197:50051: connect: connection refused" 2025-12-04T11:55:48.790506372Z W1204 11:55:48.790407 1 logging.go:55] [core] [Channel #2 SubChannel #5]grpc: addrConn.createTransport failed to connect to {Addr: "community-operators.openshift-marketplace.svc:50051", ServerName: "community-operators.openshift-marketplace.svc:50051", }. Err: connection error: desc = "transport: Error while dialing: dial tcp 172.30.111.58:50051: connect: connection refused" 2025-12-04T11:55:51.365600934Z W1204 11:55:51.365488 1 logging.go:55] [core] [Channel #1 SubChannel #3]grpc: addrConn.createTransport failed to connect to {Addr: "certified-operators.openshift-marketplace.svc:50051", ServerName: "certified-operators.openshift-marketplace.svc:50051", }. Err: connection error: desc = "transport: Error while dialing: dial tcp 172.30.138.0:50051: connect: connection refused" 2025-12-04T11:55:59.980860951Z W1204 11:55:59.980633 1 logging.go:55] [core] [Channel #4 SubChannel #8]grpc: addrConn.createTransport failed to connect to {Addr: "redhat-marketplace.openshift-marketplace.svc:50051", ServerName: "redhat-marketplace.openshift-marketplace.svc:50051", }. Err: connection error: desc = "transport: Error while dialing: dial tcp 172.30.165.235:50051: connect: connection refused" 2025-12-04T11:56:10.992859832Z W1204 11:56:10.992702 1 logging.go:55] [core] [Channel #6 SubChannel #7]grpc: addrConn.createTransport failed to connect to {Addr: "redhat-operators.openshift-marketplace.svc:50051", ServerName: "redhat-operators.openshift-marketplace.svc:50051", }. Err: connection error: desc = "transport: Error while dialing: dial tcp 172.30.86.197:50051: connect: connection refused" 2025-12-04T11:56:11.586657014Z time="2025-12-04T11:56:11Z" level=info msg="updating PackageManifest based on CatalogSource changes: {redhat-operators openshift-marketplace}" action="sync catalogsource" address="redhat-operators.openshift-marketplace.svc:50051" name=redhat-operators namespace=openshift-marketplace 2025-12-04T11:56:11.586657014Z time="2025-12-04T11:56:11Z" level=warning msg="error getting bundle stream" action="refresh cache" err="rpc error: code = Unavailable desc = connection error: desc = \"transport: Error while dialing: dial tcp 172.30.86.197:50051: connect: connection refused\"" source="{redhat-operators openshift-marketplace}" 2025-12-04T11:56:11.590272374Z time="2025-12-04T11:56:11Z" level=info msg="updating PackageManifest based on CatalogSource changes: {community-operators openshift-marketplace}" action="sync catalogsource" address="community-operators.openshift-marketplace.svc:50051" name=community-operators namespace=openshift-marketplace 2025-12-04T11:56:11.590272374Z time="2025-12-04T11:56:11Z" level=warning msg="error getting bundle stream" action="refresh cache" err="rpc error: code = Unavailable desc = connection error: desc = \"transport: Error while dialing: dial tcp 172.30.111.58:50051: connect: connection refused\"" source="{community-operators openshift-marketplace}" 2025-12-04T11:56:12.015174328Z time="2025-12-04T11:56:12Z" level=info msg="updating PackageManifest based on CatalogSource changes: {redhat-marketplace openshift-marketplace}" action="sync catalogsource" address="redhat-marketplace.openshift-marketplace.svc:50051" name=redhat-marketplace namespace=openshift-marketplace 2025-12-04T11:56:12.015238310Z time="2025-12-04T11:56:12Z" level=warning msg="error getting bundle stream" action="refresh cache" err="rpc error: code = Unavailable desc = connection error: desc = \"transport: Error while dialing: dial tcp 172.30.165.235:50051: connect: connection refused\"" source="{redhat-marketplace openshift-marketplace}" 2025-12-04T11:56:12.022112079Z time="2025-12-04T11:56:12Z" level=info msg="updating PackageManifest based on CatalogSource changes: {certified-operators openshift-marketplace}" action="sync catalogsource" address="certified-operators.openshift-marketplace.svc:50051" name=certified-operators namespace=openshift-marketplace 2025-12-04T11:56:12.022157550Z time="2025-12-04T11:56:12Z" level=warning msg="error getting bundle stream" action="refresh cache" err="rpc error: code = Unavailable desc = connection error: desc = \"transport: Error while dialing: dial tcp 172.30.138.0:50051: connect: connection refused\"" source="{certified-operators openshift-marketplace}" 2025-12-04T11:57:33.931544641Z W1204 11:57:33.931385 1 logging.go:55] [core] [Channel #1 SubChannel #3]grpc: addrConn.createTransport failed to connect to {Addr: "certified-operators.openshift-marketplace.svc:50051", ServerName: "certified-operators.openshift-marketplace.svc:50051", }. Err: connection error: desc = "transport: Error while dialing: dial tcp 172.30.138.0:50051: connect: connection refused" 2025-12-04T11:57:38.015610959Z W1204 11:57:38.015516 1 logging.go:55] [core] [Channel #4 SubChannel #8]grpc: addrConn.createTransport failed to connect to {Addr: "redhat-marketplace.openshift-marketplace.svc:50051", ServerName: "redhat-marketplace.openshift-marketplace.svc:50051", }. Err: connection error: desc = "transport: Error while dialing: dial tcp 172.30.165.235:50051: connect: connection refused" 2025-12-04T11:58:06.737430949Z W1204 11:58:06.737329 1 logging.go:55] [core] [Channel #6 SubChannel #7]grpc: addrConn.createTransport failed to connect to {Addr: "redhat-operators.openshift-marketplace.svc:50051", ServerName: "redhat-operators.openshift-marketplace.svc:50051", }. Err: connection error: desc = "transport: Error while dialing: dial tcp 172.30.86.197:50051: connect: connection refused" 2025-12-04T11:58:10.622895020Z W1204 11:58:10.622687 1 logging.go:55] [core] [Channel #2 SubChannel #5]grpc: addrConn.createTransport failed to connect to {Addr: "community-operators.openshift-marketplace.svc:50051", ServerName: "community-operators.openshift-marketplace.svc:50051", }. Err: connection error: desc = "transport: Error while dialing: dial tcp 172.30.111.58:50051: connect: connection refused" 2025-12-04T11:59:35.899861985Z W1204 11:59:35.899761 1 logging.go:55] [core] [Channel #4 SubChannel #8]grpc: addrConn.createTransport failed to connect to {Addr: "redhat-marketplace.openshift-marketplace.svc:50051", ServerName: "redhat-marketplace.openshift-marketplace.svc:50051", }. Err: connection error: desc = "transport: Error while dialing: dial tcp 172.30.165.235:50051: connect: connection refused" 2025-12-04T11:59:49.203564299Z W1204 11:59:49.203477 1 logging.go:55] [core] [Channel #1 SubChannel #3]grpc: addrConn.createTransport failed to connect to {Addr: "certified-operators.openshift-marketplace.svc:50051", ServerName: "certified-operators.openshift-marketplace.svc:50051", }. Err: connection error: desc = "transport: Error while dialing: dial tcp 172.30.138.0:50051: connect: connection refused" 2025-12-04T11:59:56.777380678Z time="2025-12-04T11:59:56Z" level=info msg="updating PackageManifest based on CatalogSource changes: {certified-operators openshift-marketplace}" action="sync catalogsource" address="certified-operators.openshift-marketplace.svc:50051" name=certified-operators namespace=openshift-marketplace 2025-12-04T11:59:56.777380678Z time="2025-12-04T11:59:56Z" level=info msg="updating PackageManifest based on CatalogSource changes: {community-operators openshift-marketplace}" action="sync catalogsource" address="community-operators.openshift-marketplace.svc:50051" name=community-operators namespace=openshift-marketplace 2025-12-04T11:59:56.777480891Z time="2025-12-04T11:59:56Z" level=warning msg="error getting bundle stream" action="refresh cache" err="rpc error: code = Unavailable desc = connection error: desc = \"transport: Error while dialing: dial tcp 172.30.111.58:50051: connect: connection refused\"" source="{community-operators openshift-marketplace}" 2025-12-04T11:59:56.777520182Z time="2025-12-04T11:59:56Z" level=info msg="updating PackageManifest based on CatalogSource changes: {redhat-marketplace openshift-marketplace}" action="sync catalogsource" address="redhat-marketplace.openshift-marketplace.svc:50051" name=redhat-marketplace namespace=openshift-marketplace 2025-12-04T11:59:56.777550633Z time="2025-12-04T11:59:56Z" level=warning msg="error getting bundle stream" action="refresh cache" err="rpc error: code = Unavailable desc = connection error: desc = \"transport: Error while dialing: dial tcp 172.30.165.235:50051: connect: connection refused\"" source="{redhat-marketplace openshift-marketplace}" 2025-12-04T11:59:56.777601195Z time="2025-12-04T11:59:56Z" level=info msg="updating PackageManifest based on CatalogSource changes: {redhat-operators openshift-marketplace}" action="sync catalogsource" address="redhat-operators.openshift-marketplace.svc:50051" name=redhat-operators namespace=openshift-marketplace 2025-12-04T11:59:56.777601195Z time="2025-12-04T11:59:56Z" level=warning msg="error getting bundle stream" action="refresh cache" err="rpc error: code = Unavailable desc = connection error: desc = \"transport: Error while dialing: dial tcp 172.30.138.0:50051: connect: connection refused\"" source="{certified-operators openshift-marketplace}" 2025-12-04T11:59:56.777637586Z time="2025-12-04T11:59:56Z" level=warning msg="error getting bundle stream" action="refresh cache" err="rpc error: code = Unavailable desc = connection error: desc = \"transport: Error while dialing: dial tcp 172.30.86.197:50051: connect: connection refused\"" source="{redhat-operators openshift-marketplace}" 2025-12-04T12:00:09.492846791Z W1204 12:00:09.492749 1 logging.go:55] [core] [Channel #2 SubChannel #5]grpc: addrConn.createTransport failed to connect to {Addr: "community-operators.openshift-marketplace.svc:50051", ServerName: "community-operators.openshift-marketplace.svc:50051", }. Err: connection error: desc = "transport: Error while dialing: dial tcp 172.30.111.58:50051: connect: connection refused" 2025-12-04T12:00:14.914706137Z W1204 12:00:14.914615 1 logging.go:55] [core] [Channel #6 SubChannel #7]grpc: addrConn.createTransport failed to connect to {Addr: "redhat-operators.openshift-marketplace.svc:50051", ServerName: "redhat-operators.openshift-marketplace.svc:50051", }. Err: connection error: desc = "transport: Error while dialing: dial tcp 172.30.86.197:50051: connect: connection refused" 2025-12-04T12:01:31.043049593Z W1204 12:01:31.042963 1 logging.go:55] [core] [Channel #4 SubChannel #8]grpc: addrConn.createTransport failed to connect to {Addr: "redhat-marketplace.openshift-marketplace.svc:50051", ServerName: "redhat-marketplace.openshift-marketplace.svc:50051", }. Err: connection error: desc = "transport: Error while dialing: dial tcp 172.30.165.235:50051: connect: connection refused" 2025-12-04T12:01:45.833777575Z W1204 12:01:45.833691 1 logging.go:55] [core] [Channel #2 SubChannel #5]grpc: addrConn.createTransport failed to connect to {Addr: "community-operators.openshift-marketplace.svc:50051", ServerName: "community-operators.openshift-marketplace.svc:50051", }. Err: connection error: desc = "transport: Error while dialing: dial tcp 172.30.111.58:50051: connect: connection refused" 2025-12-04T12:01:51.803258520Z W1204 12:01:51.803199 1 logging.go:55] [core] [Channel #1 SubChannel #3]grpc: addrConn.createTransport failed to connect to {Addr: "certified-operators.openshift-marketplace.svc:50051", ServerName: "certified-operators.openshift-marketplace.svc:50051", }. Err: connection error: desc = "transport: Error while dialing: dial tcp 172.30.138.0:50051: connect: connection refused" package-server-manager_current.log0000640000000000000000000004127415114274105036634 0ustar0000000000000000config/pod/openshift-operator-lifecycle-manager/logs/package-server-manager-67477646d4-7hndf2025-12-04T11:53:58.147338565Z 2025-12-04T11:53:58Z ERROR setup unable to get cluster infrastructure status, using HA cluster values for leader election {"error": "Get \"https://172.30.0.1:443/apis/config.openshift.io/v1/infrastructures/cluster\": context deadline exceeded"} 2025-12-04T11:53:58.147338565Z github.com/openshift/operator-framework-olm/pkg/leaderelection.getLeaderElectionConfig 2025-12-04T11:53:58.147338565Z /build/pkg/leaderelection/leaderelection.go:61 2025-12-04T11:53:58.147338565Z github.com/openshift/operator-framework-olm/pkg/leaderelection.GetLeaderElectionConfig 2025-12-04T11:53:58.147338565Z /build/pkg/leaderelection/leaderelection.go:50 2025-12-04T11:53:58.147338565Z main.run 2025-12-04T11:53:58.147338565Z /build/cmd/package-server-manager/main.go:83 2025-12-04T11:53:58.147338565Z github.com/spf13/cobra.(*Command).execute 2025-12-04T11:53:58.147338565Z /build/vendor/github.com/spf13/cobra/command.go:985 2025-12-04T11:53:58.147338565Z github.com/spf13/cobra.(*Command).ExecuteC 2025-12-04T11:53:58.147338565Z /build/vendor/github.com/spf13/cobra/command.go:1117 2025-12-04T11:53:58.147338565Z github.com/spf13/cobra.(*Command).Execute 2025-12-04T11:53:58.147338565Z /build/vendor/github.com/spf13/cobra/command.go:1041 2025-12-04T11:53:58.147338565Z main.main 2025-12-04T11:53:58.147338565Z /build/cmd/package-server-manager/main.go:43 2025-12-04T11:53:58.147338565Z runtime.main 2025-12-04T11:53:58.147338565Z /usr/lib/golang/src/runtime/proc.go:271 2025-12-04T11:53:58.184872834Z 2025-12-04T11:53:58Z INFO setup starting manager 2025-12-04T11:53:58.185741112Z 2025-12-04T11:53:58Z INFO controller-runtime.metrics Starting metrics server 2025-12-04T11:53:58.185845445Z 2025-12-04T11:53:58Z INFO starting server {"name": "pprof", "addr": "[::]:6060"} 2025-12-04T11:53:58.187253430Z 2025-12-04T11:53:58Z INFO starting server {"name": "health probe", "addr": "[::]:8080"} 2025-12-04T11:53:58.187442366Z 2025-12-04T11:53:58Z INFO controller-runtime.metrics Serving metrics server {"bindAddress": ":9090", "secure": false} 2025-12-04T11:53:58.187451366Z I1204 11:53:58.187418 1 leaderelection.go:254] attempting to acquire leader lease openshift-operator-lifecycle-manager/packageserver-controller-lock... 2025-12-04T11:54:58.193660170Z E1204 11:54:58.193570 1 leaderelection.go:436] error retrieving resource lock openshift-operator-lifecycle-manager/packageserver-controller-lock: the server was unable to return a response in the time allotted, but may still be processing the request (get leases.coordination.k8s.io packageserver-controller-lock) 2025-12-04T11:58:24.357638040Z I1204 11:58:24.357538 1 leaderelection.go:268] successfully acquired lease openshift-operator-lifecycle-manager/packageserver-controller-lock 2025-12-04T11:58:24.363282261Z 2025-12-04T11:58:24Z DEBUG events package-server-manager-67477646d4-7hndf_efa69d16-9250-446b-9dbd-9d804226ad9f became leader {"type": "Normal", "object": {"kind":"Lease","namespace":"openshift-operator-lifecycle-manager","name":"packageserver-controller-lock","uid":"79235196-994a-4023-9bf8-240670cd2300","apiVersion":"coordination.k8s.io/v1","resourceVersion":"11575"}, "reason": "LeaderElection"} 2025-12-04T11:58:24.363335842Z 2025-12-04T11:58:24Z INFO Starting EventSource {"controller": "clusterserviceversion", "controllerGroup": "operators.coreos.com", "controllerKind": "ClusterServiceVersion", "source": "kind source: *v1alpha1.ClusterServiceVersion"} 2025-12-04T11:58:24.363372763Z 2025-12-04T11:58:24Z INFO Starting EventSource {"controller": "clusterserviceversion", "controllerGroup": "operators.coreos.com", "controllerKind": "ClusterServiceVersion", "source": "kind source: *v1.Infrastructure"} 2025-12-04T11:58:24.363372763Z 2025-12-04T11:58:24Z INFO Starting Controller {"controller": "clusterserviceversion", "controllerGroup": "operators.coreos.com", "controllerKind": "ClusterServiceVersion"} 2025-12-04T11:58:24.494031945Z 2025-12-04T11:58:24Z INFO Starting workers {"controller": "clusterserviceversion", "controllerGroup": "operators.coreos.com", "controllerKind": "ClusterServiceVersion", "worker count": 1} 2025-12-04T11:58:24.494031945Z 2025-12-04T11:58:24Z INFO controllers.packageserver requeueing the packageserver deployment after encountering infrastructure event {"infrastructure": "cluster"} 2025-12-04T11:58:24.498243847Z 2025-12-04T11:58:24Z INFO controllers.packageserver handling current request {"csv": {"name":"packageserver","namespace":"openshift-operator-lifecycle-manager"}, "request": "openshift-operator-lifecycle-manager/packageserver"} 2025-12-04T11:58:24.498243847Z 2025-12-04T11:58:24Z INFO controllers.packageserver checking to see if required RBAC exists {"csv": {"name":"packageserver","namespace":"openshift-operator-lifecycle-manager"}} 2025-12-04T11:58:24.802731785Z 2025-12-04T11:58:24Z INFO controllers.packageserver confimed required RBAC exists {"csv": {"name":"packageserver","namespace":"openshift-operator-lifecycle-manager"}} 2025-12-04T11:58:24.802842248Z 2025-12-04T11:58:24Z INFO controllers.packageserver currently topology mode {"csv": {"name":"packageserver","namespace":"openshift-operator-lifecycle-manager"}, "highly available": false} 2025-12-04T11:58:24.846354951Z 2025-12-04T11:58:24Z INFO controllers.packageserver reconciliation result {"csv": {"name":"packageserver","namespace":"openshift-operator-lifecycle-manager"}, "res": "unchanged"} 2025-12-04T11:58:24.846428983Z 2025-12-04T11:58:24Z INFO controllers.packageserver finished request reconciliation {"csv": {"name":"packageserver","namespace":"openshift-operator-lifecycle-manager"}} 2025-12-04T11:58:24.846554146Z 2025-12-04T11:58:24Z INFO controllers.packageserver handling current request {"csv": {"name":"packageserver","namespace":"openshift-operator-lifecycle-manager"}, "request": "openshift-operator-lifecycle-manager/packageserver"} 2025-12-04T11:58:24.846554146Z 2025-12-04T11:58:24Z INFO controllers.packageserver checking to see if required RBAC exists {"csv": {"name":"packageserver","namespace":"openshift-operator-lifecycle-manager"}} 2025-12-04T11:58:24.846723771Z 2025-12-04T11:58:24Z INFO controllers.packageserver confimed required RBAC exists {"csv": {"name":"packageserver","namespace":"openshift-operator-lifecycle-manager"}} 2025-12-04T11:58:24.846757702Z 2025-12-04T11:58:24Z INFO controllers.packageserver currently topology mode {"csv": {"name":"packageserver","namespace":"openshift-operator-lifecycle-manager"}, "highly available": false} 2025-12-04T11:58:24.861993728Z 2025-12-04T11:58:24Z INFO controllers.packageserver reconciliation result {"csv": {"name":"packageserver","namespace":"openshift-operator-lifecycle-manager"}, "res": "unchanged"} 2025-12-04T11:58:24.861993728Z 2025-12-04T11:58:24Z INFO controllers.packageserver finished request reconciliation {"csv": {"name":"packageserver","namespace":"openshift-operator-lifecycle-manager"}} 2025-12-04T11:59:16.622352237Z E1204 11:59:16.622225 1 leaderelection.go:429] Failed to update lock optimitically: Put "https://172.30.0.1:443/apis/coordination.k8s.io/v1/namespaces/openshift-operator-lifecycle-manager/leases/packageserver-controller-lock": dial tcp 172.30.0.1:443: connect: connection refused, falling back to slow path 2025-12-04T11:59:16.626313792Z E1204 11:59:16.626195 1 leaderelection.go:436] error retrieving resource lock openshift-operator-lifecycle-manager/packageserver-controller-lock: Get "https://172.30.0.1:443/apis/coordination.k8s.io/v1/namespaces/openshift-operator-lifecycle-manager/leases/packageserver-controller-lock": dial tcp 172.30.0.1:443: connect: connection refused 2025-12-04T11:59:47.172427799Z 2025-12-04T11:59:47Z INFO controllers.packageserver handling current request {"csv": {"name":"packageserver","namespace":"openshift-operator-lifecycle-manager"}, "request": "openshift-operator-lifecycle-manager/packageserver"} 2025-12-04T11:59:47.172427799Z 2025-12-04T11:59:47Z INFO controllers.packageserver checking to see if required RBAC exists {"csv": {"name":"packageserver","namespace":"openshift-operator-lifecycle-manager"}} 2025-12-04T11:59:47.172427799Z 2025-12-04T11:59:47Z INFO controllers.packageserver confimed required RBAC exists {"csv": {"name":"packageserver","namespace":"openshift-operator-lifecycle-manager"}} 2025-12-04T11:59:47.172427799Z 2025-12-04T11:59:47Z INFO controllers.packageserver currently topology mode {"csv": {"name":"packageserver","namespace":"openshift-operator-lifecycle-manager"}, "highly available": false} 2025-12-04T11:59:47.188734547Z 2025-12-04T11:59:47Z INFO controllers.packageserver reconciliation result {"csv": {"name":"packageserver","namespace":"openshift-operator-lifecycle-manager"}, "res": "unchanged"} 2025-12-04T11:59:47.188734547Z 2025-12-04T11:59:47Z INFO controllers.packageserver finished request reconciliation {"csv": {"name":"packageserver","namespace":"openshift-operator-lifecycle-manager"}} 2025-12-04T11:59:58.953983434Z 2025-12-04T11:59:58Z INFO controllers.packageserver requeueing the packageserver deployment after encountering infrastructure event {"infrastructure": "cluster"} 2025-12-04T11:59:58.954666676Z 2025-12-04T11:59:58Z INFO controllers.packageserver requeueing the packageserver deployment after encountering infrastructure event {"infrastructure": "cluster"} 2025-12-04T11:59:58.956992449Z 2025-12-04T11:59:58Z INFO controllers.packageserver handling current request {"csv": {"name":"packageserver","namespace":"openshift-operator-lifecycle-manager"}, "request": "openshift-operator-lifecycle-manager/packageserver"} 2025-12-04T11:59:58.956992449Z 2025-12-04T11:59:58Z INFO controllers.packageserver checking to see if required RBAC exists {"csv": {"name":"packageserver","namespace":"openshift-operator-lifecycle-manager"}} 2025-12-04T11:59:58.957122093Z 2025-12-04T11:59:58Z INFO controllers.packageserver confimed required RBAC exists {"csv": {"name":"packageserver","namespace":"openshift-operator-lifecycle-manager"}} 2025-12-04T11:59:58.957173785Z 2025-12-04T11:59:58Z INFO controllers.packageserver currently topology mode {"csv": {"name":"packageserver","namespace":"openshift-operator-lifecycle-manager"}, "highly available": false} 2025-12-04T11:59:58.977295842Z 2025-12-04T11:59:58Z INFO controllers.packageserver reconciliation result {"csv": {"name":"packageserver","namespace":"openshift-operator-lifecycle-manager"}, "res": "unchanged"} 2025-12-04T11:59:58.977295842Z 2025-12-04T11:59:58Z INFO controllers.packageserver finished request reconciliation {"csv": {"name":"packageserver","namespace":"openshift-operator-lifecycle-manager"}} 2025-12-04T12:00:59.135014857Z 2025-12-04T12:00:59Z INFO controllers.packageserver handling current request {"csv": {"name":"packageserver","namespace":"openshift-operator-lifecycle-manager"}, "request": "openshift-operator-lifecycle-manager/packageserver"} 2025-12-04T12:00:59.135870585Z 2025-12-04T12:00:59Z INFO controllers.packageserver checking to see if required RBAC exists {"csv": {"name":"packageserver","namespace":"openshift-operator-lifecycle-manager"}} 2025-12-04T12:00:59.136334520Z 2025-12-04T12:00:59Z INFO controllers.packageserver confimed required RBAC exists {"csv": {"name":"packageserver","namespace":"openshift-operator-lifecycle-manager"}} 2025-12-04T12:00:59.136459744Z 2025-12-04T12:00:59Z INFO controllers.packageserver currently topology mode {"csv": {"name":"packageserver","namespace":"openshift-operator-lifecycle-manager"}, "highly available": false} 2025-12-04T12:00:59.151870561Z 2025-12-04T12:00:59Z INFO controllers.packageserver reconciliation result {"csv": {"name":"packageserver","namespace":"openshift-operator-lifecycle-manager"}, "res": "unchanged"} 2025-12-04T12:00:59.151914733Z 2025-12-04T12:00:59Z INFO controllers.packageserver finished request reconciliation {"csv": {"name":"packageserver","namespace":"openshift-operator-lifecycle-manager"}} 2025-12-04T12:00:59.382256878Z 2025-12-04T12:00:59Z INFO controllers.packageserver handling current request {"csv": {"name":"packageserver","namespace":"openshift-operator-lifecycle-manager"}, "request": "openshift-operator-lifecycle-manager/packageserver"} 2025-12-04T12:00:59.382256878Z 2025-12-04T12:00:59Z INFO controllers.packageserver checking to see if required RBAC exists {"csv": {"name":"packageserver","namespace":"openshift-operator-lifecycle-manager"}} 2025-12-04T12:00:59.382322100Z 2025-12-04T12:00:59Z INFO controllers.packageserver confimed required RBAC exists {"csv": {"name":"packageserver","namespace":"openshift-operator-lifecycle-manager"}} 2025-12-04T12:00:59.382322100Z 2025-12-04T12:00:59Z INFO controllers.packageserver currently topology mode {"csv": {"name":"packageserver","namespace":"openshift-operator-lifecycle-manager"}, "highly available": false} 2025-12-04T12:00:59.397986557Z 2025-12-04T12:00:59Z INFO controllers.packageserver reconciliation result {"csv": {"name":"packageserver","namespace":"openshift-operator-lifecycle-manager"}, "res": "unchanged"} 2025-12-04T12:00:59.397986557Z 2025-12-04T12:00:59Z INFO controllers.packageserver finished request reconciliation {"csv": {"name":"packageserver","namespace":"openshift-operator-lifecycle-manager"}} 2025-12-04T12:00:59.477976580Z 2025-12-04T12:00:59Z INFO controllers.packageserver handling current request {"csv": {"name":"packageserver","namespace":"openshift-operator-lifecycle-manager"}, "request": "openshift-operator-lifecycle-manager/packageserver"} 2025-12-04T12:00:59.477976580Z 2025-12-04T12:00:59Z INFO controllers.packageserver checking to see if required RBAC exists {"csv": {"name":"packageserver","namespace":"openshift-operator-lifecycle-manager"}} 2025-12-04T12:00:59.478034562Z 2025-12-04T12:00:59Z INFO controllers.packageserver confimed required RBAC exists {"csv": {"name":"packageserver","namespace":"openshift-operator-lifecycle-manager"}} 2025-12-04T12:00:59.478057642Z 2025-12-04T12:00:59Z INFO controllers.packageserver currently topology mode {"csv": {"name":"packageserver","namespace":"openshift-operator-lifecycle-manager"}, "highly available": false} 2025-12-04T12:00:59.509059084Z 2025-12-04T12:00:59Z INFO controllers.packageserver reconciliation result {"csv": {"name":"packageserver","namespace":"openshift-operator-lifecycle-manager"}, "res": "unchanged"} 2025-12-04T12:00:59.509059084Z 2025-12-04T12:00:59Z INFO controllers.packageserver finished request reconciliation {"csv": {"name":"packageserver","namespace":"openshift-operator-lifecycle-manager"}} 2025-12-04T12:00:59.686016129Z 2025-12-04T12:00:59Z INFO controllers.packageserver handling current request {"csv": {"name":"packageserver","namespace":"openshift-operator-lifecycle-manager"}, "request": "openshift-operator-lifecycle-manager/packageserver"} 2025-12-04T12:00:59.686016129Z 2025-12-04T12:00:59Z INFO controllers.packageserver checking to see if required RBAC exists {"csv": {"name":"packageserver","namespace":"openshift-operator-lifecycle-manager"}} 2025-12-04T12:00:59.686076641Z 2025-12-04T12:00:59Z INFO controllers.packageserver confimed required RBAC exists {"csv": {"name":"packageserver","namespace":"openshift-operator-lifecycle-manager"}} 2025-12-04T12:00:59.686139183Z 2025-12-04T12:00:59Z INFO controllers.packageserver currently topology mode {"csv": {"name":"packageserver","namespace":"openshift-operator-lifecycle-manager"}, "highly available": false} 2025-12-04T12:00:59.722983229Z 2025-12-04T12:00:59Z INFO controllers.packageserver reconciliation result {"csv": {"name":"packageserver","namespace":"openshift-operator-lifecycle-manager"}, "res": "unchanged"} 2025-12-04T12:00:59.722983229Z 2025-12-04T12:00:59Z INFO controllers.packageserver finished request reconciliation {"csv": {"name":"packageserver","namespace":"openshift-operator-lifecycle-manager"}} 2025-12-04T12:00:59.835221004Z 2025-12-04T12:00:59Z INFO controllers.packageserver handling current request {"csv": {"name":"packageserver","namespace":"openshift-operator-lifecycle-manager"}, "request": "openshift-operator-lifecycle-manager/packageserver"} 2025-12-04T12:00:59.835221004Z 2025-12-04T12:00:59Z INFO controllers.packageserver checking to see if required RBAC exists {"csv": {"name":"packageserver","namespace":"openshift-operator-lifecycle-manager"}} 2025-12-04T12:00:59.835355478Z 2025-12-04T12:00:59Z INFO controllers.packageserver confimed required RBAC exists {"csv": {"name":"packageserver","namespace":"openshift-operator-lifecycle-manager"}} 2025-12-04T12:00:59.835387259Z 2025-12-04T12:00:59Z INFO controllers.packageserver currently topology mode {"csv": {"name":"packageserver","namespace":"openshift-operator-lifecycle-manager"}, "highly available": false} 2025-12-04T12:00:59.855425334Z 2025-12-04T12:00:59Z INFO controllers.packageserver reconciliation result {"csv": {"name":"packageserver","namespace":"openshift-operator-lifecycle-manager"}, "res": "unchanged"} 2025-12-04T12:00:59.855425334Z 2025-12-04T12:00:59Z INFO controllers.packageserver finished request reconciliation {"csv": {"name":"packageserver","namespace":"openshift-operator-lifecycle-manager"}} kube-rbac-proxy_current.log0000640000000000000000000000160415114274105035330 0ustar0000000000000000config/pod/openshift-operator-lifecycle-manager/logs/package-server-manager-67477646d4-7hndf2025-12-04T11:38:20.258906139Z W1204 11:38:20.258600 1 deprecated.go:66] 2025-12-04T11:38:20.258906139Z ==== Removed Flag Warning ====================== 2025-12-04T11:38:20.258906139Z 2025-12-04T11:38:20.258906139Z logtostderr is removed in the k8s upstream and has no effect any more. 2025-12-04T11:38:20.258906139Z 2025-12-04T11:38:20.258906139Z =============================================== 2025-12-04T11:38:20.258906139Z 2025-12-04T11:38:20.259409694Z I1204 11:38:20.259378 1 kube-rbac-proxy.go:233] Valid token audiences: 2025-12-04T11:38:20.260784739Z I1204 11:38:20.260752 1 kube-rbac-proxy.go:347] Reading certificate files 2025-12-04T11:38:20.261203263Z I1204 11:38:20.261151 1 kube-rbac-proxy.go:395] Starting TCP socket on 0.0.0.0:8443 2025-12-04T11:38:20.262742322Z I1204 11:38:20.261769 1 kube-rbac-proxy.go:402] Listening securely on 0.0.0.0:8443 package-server-manager_previous.log0000640000000000000000000001665715114274105037035 0ustar0000000000000000config/pod/openshift-operator-lifecycle-manager/logs/package-server-manager-67477646d4-7hndf2025-12-04T11:44:31.633515913Z 2025-12-04T11:44:31Z ERROR setup unable to get cluster infrastructure status, using HA cluster values for leader election {"error": "Get \"https://172.30.0.1:443/apis/config.openshift.io/v1/infrastructures/cluster\": context deadline exceeded"} 2025-12-04T11:44:31.633515913Z github.com/openshift/operator-framework-olm/pkg/leaderelection.getLeaderElectionConfig 2025-12-04T11:44:31.633515913Z /build/pkg/leaderelection/leaderelection.go:61 2025-12-04T11:44:31.633515913Z github.com/openshift/operator-framework-olm/pkg/leaderelection.GetLeaderElectionConfig 2025-12-04T11:44:31.633515913Z /build/pkg/leaderelection/leaderelection.go:50 2025-12-04T11:44:31.633515913Z main.run 2025-12-04T11:44:31.633515913Z /build/cmd/package-server-manager/main.go:83 2025-12-04T11:44:31.633515913Z github.com/spf13/cobra.(*Command).execute 2025-12-04T11:44:31.633515913Z /build/vendor/github.com/spf13/cobra/command.go:985 2025-12-04T11:44:31.633515913Z github.com/spf13/cobra.(*Command).ExecuteC 2025-12-04T11:44:31.633515913Z /build/vendor/github.com/spf13/cobra/command.go:1117 2025-12-04T11:44:31.633515913Z github.com/spf13/cobra.(*Command).Execute 2025-12-04T11:44:31.633515913Z /build/vendor/github.com/spf13/cobra/command.go:1041 2025-12-04T11:44:31.633515913Z main.main 2025-12-04T11:44:31.633515913Z /build/cmd/package-server-manager/main.go:43 2025-12-04T11:44:31.633515913Z runtime.main 2025-12-04T11:44:31.633515913Z /usr/lib/golang/src/runtime/proc.go:271 2025-12-04T11:44:31.671936005Z 2025-12-04T11:44:31Z INFO setup starting manager 2025-12-04T11:44:31.673760289Z 2025-12-04T11:44:31Z INFO controller-runtime.metrics Starting metrics server 2025-12-04T11:44:31.674059220Z 2025-12-04T11:44:31Z INFO starting server {"name": "pprof", "addr": "[::]:6060"} 2025-12-04T11:44:31.674183124Z 2025-12-04T11:44:31Z INFO controller-runtime.metrics Serving metrics server {"bindAddress": ":9090", "secure": false} 2025-12-04T11:44:31.674183124Z 2025-12-04T11:44:31Z INFO starting server {"name": "health probe", "addr": "[::]:8080"} 2025-12-04T11:44:31.675409118Z I1204 11:44:31.675346 1 leaderelection.go:254] attempting to acquire leader lease openshift-operator-lifecycle-manager/packageserver-controller-lock... 2025-12-04T11:49:29.564283335Z I1204 11:49:29.564129 1 leaderelection.go:268] successfully acquired lease openshift-operator-lifecycle-manager/packageserver-controller-lock 2025-12-04T11:49:29.566144633Z 2025-12-04T11:49:29Z DEBUG events package-server-manager-67477646d4-7hndf_5a767bc3-9a75-405d-8427-31679e2dfc65 became leader {"type": "Normal", "object": {"kind":"Lease","namespace":"openshift-operator-lifecycle-manager","name":"packageserver-controller-lock","uid":"79235196-994a-4023-9bf8-240670cd2300","apiVersion":"coordination.k8s.io/v1","resourceVersion":"10708"}, "reason": "LeaderElection"} 2025-12-04T11:49:29.567343950Z 2025-12-04T11:49:29Z INFO Starting EventSource {"controller": "clusterserviceversion", "controllerGroup": "operators.coreos.com", "controllerKind": "ClusterServiceVersion", "source": "kind source: *v1alpha1.ClusterServiceVersion"} 2025-12-04T11:49:29.567458554Z 2025-12-04T11:49:29Z INFO Starting EventSource {"controller": "clusterserviceversion", "controllerGroup": "operators.coreos.com", "controllerKind": "ClusterServiceVersion", "source": "kind source: *v1.Infrastructure"} 2025-12-04T11:49:29.567458554Z 2025-12-04T11:49:29Z INFO Starting Controller {"controller": "clusterserviceversion", "controllerGroup": "operators.coreos.com", "controllerKind": "ClusterServiceVersion"} 2025-12-04T11:49:29.699260509Z 2025-12-04T11:49:29Z INFO controllers.packageserver requeueing the packageserver deployment after encountering infrastructure event {"infrastructure": "cluster"} 2025-12-04T11:49:29.699260509Z 2025-12-04T11:49:29Z INFO Starting workers {"controller": "clusterserviceversion", "controllerGroup": "operators.coreos.com", "controllerKind": "ClusterServiceVersion", "worker count": 1} 2025-12-04T11:49:29.699727374Z 2025-12-04T11:49:29Z INFO controllers.packageserver handling current request {"csv": {"name":"packageserver","namespace":"openshift-operator-lifecycle-manager"}, "request": "openshift-operator-lifecycle-manager/packageserver"} 2025-12-04T11:49:29.699768265Z 2025-12-04T11:49:29Z INFO controllers.packageserver checking to see if required RBAC exists {"csv": {"name":"packageserver","namespace":"openshift-operator-lifecycle-manager"}} 2025-12-04T11:49:30.003992014Z 2025-12-04T11:49:30Z INFO controllers.packageserver confimed required RBAC exists {"csv": {"name":"packageserver","namespace":"openshift-operator-lifecycle-manager"}} 2025-12-04T11:49:30.004049776Z 2025-12-04T11:49:30Z INFO controllers.packageserver currently topology mode {"csv": {"name":"packageserver","namespace":"openshift-operator-lifecycle-manager"}, "highly available": false} 2025-12-04T11:49:30.041527143Z 2025-12-04T11:49:30Z INFO controllers.packageserver reconciliation result {"csv": {"name":"packageserver","namespace":"openshift-operator-lifecycle-manager"}, "res": "unchanged"} 2025-12-04T11:49:30.041527143Z 2025-12-04T11:49:30Z INFO controllers.packageserver finished request reconciliation {"csv": {"name":"packageserver","namespace":"openshift-operator-lifecycle-manager"}} 2025-12-04T11:51:47.640029961Z E1204 11:51:47.639922 1 leaderelection.go:429] Failed to update lock optimitically: Timeout: request did not complete within requested timeout - context deadline exceeded, falling back to slow path 2025-12-04T11:52:47.645551249Z E1204 11:52:47.645428 1 leaderelection.go:436] error retrieving resource lock openshift-operator-lifecycle-manager/packageserver-controller-lock: the server was unable to return a response in the time allotted, but may still be processing the request (get leases.coordination.k8s.io packageserver-controller-lock) 2025-12-04T11:53:00.631332340Z I1204 11:53:00.630259 1 leaderelection.go:297] failed to renew lease openshift-operator-lifecycle-manager/packageserver-controller-lock: timed out waiting for the condition 2025-12-04T11:53:34.642156263Z E1204 11:53:34.642078 1 leaderelection.go:322] Failed to release lock: Timeout: request did not complete within requested timeout - context deadline exceeded 2025-12-04T11:53:34.642483143Z 2025-12-04T11:53:34Z INFO Stopping and waiting for non leader election runnables 2025-12-04T11:53:34.642483143Z 2025-12-04T11:53:34Z INFO Stopping and waiting for leader election runnables 2025-12-04T11:53:34.642483143Z 2025-12-04T11:53:34Z ERROR setup problem running manager {"error": "leader election lost"} 2025-12-04T11:53:34.642483143Z main.run 2025-12-04T11:53:34.642483143Z /build/cmd/package-server-manager/main.go:150 2025-12-04T11:53:34.642483143Z github.com/spf13/cobra.(*Command).execute 2025-12-04T11:53:34.642483143Z /build/vendor/github.com/spf13/cobra/command.go:985 2025-12-04T11:53:34.642483143Z github.com/spf13/cobra.(*Command).ExecuteC 2025-12-04T11:53:34.642483143Z /build/vendor/github.com/spf13/cobra/command.go:1117 2025-12-04T11:53:34.642483143Z github.com/spf13/cobra.(*Command).Execute 2025-12-04T11:53:34.642483143Z /build/vendor/github.com/spf13/cobra/command.go:1041 2025-12-04T11:53:34.642483143Z main.main 2025-12-04T11:53:34.642483143Z /build/cmd/package-server-manager/main.go:43 2025-12-04T11:53:34.642483143Z runtime.main 2025-12-04T11:53:34.642483143Z /usr/lib/golang/src/runtime/proc.go:271 2025-12-04T11:53:34.642629948Z Error: leader election lost 2025-12-04T11:53:34.642629948Z encountered an error while executing the binary: leader election lost config/pod/openshift-operator-lifecycle-manager/package-server-manager-67477646d4-7hndf.json0000640000000000000000000002333615114274105030065 0ustar0000000000000000{"metadata":{"name":"package-server-manager-67477646d4-7hndf","generateName":"package-server-manager-67477646d4-","namespace":"openshift-operator-lifecycle-manager","uid":"72faf6d6-e8ca-43d1-b93e-67c11f8d3b46","resourceVersion":"11954","creationTimestamp":"2025-12-04T11:35:50Z","labels":{"app":"package-server-manager","pod-template-hash":"67477646d4"},"annotations":{"k8s.ovn.org/pod-networks":"{\"default\":{\"ip_addresses\":[\"10.128.0.13/23\"],\"mac_address\":\"0a:58:0a:80:00:0d\",\"gateway_ips\":[\"10.128.0.1\"],\"routes\":[{\"dest\":\"10.128.0.0/16\",\"nextHop\":\"10.128.0.1\"},{\"dest\":\"172.30.0.0/16\",\"nextHop\":\"10.128.0.1\"},{\"dest\":\"169.254.0.5/32\",\"nextHop\":\"10.128.0.1\"},{\"dest\":\"100.64.0.0/16\",\"nextHop\":\"10.128.0.1\"}],\"ip_address\":\"10.128.0.13/23\",\"gateway_ip\":\"10.128.0.1\",\"role\":\"primary\"}}","k8s.v1.cni.cncf.io/network-status":"[{\n \"name\": \"ovn-kubernetes\",\n \"interface\": \"eth0\",\n \"ips\": [\n \"10.128.0.13\"\n ],\n \"mac\": \"0a:58:0a:80:00:0d\",\n \"default\": true,\n \"dns\": {}\n}]","openshift.io/required-scc":"restricted-v2","openshift.io/scc":"restricted-v2","seccomp.security.alpha.kubernetes.io/pod":"runtime/default"},"ownerReferences":[{"apiVersion":"apps/v1","kind":"ReplicaSet","name":"package-server-manager-67477646d4","uid":"b300c2d5-fa0d-4c7e-937b-2a9397768709","controller":true,"blockOwnerDeletion":true}]},"spec":{"volumes":[{"name":"package-server-manager-serving-cert","secret":{"secretName":"package-server-manager-serving-cert","defaultMode":420}},{"name":"kube-api-access-ckwnn","projected":{"sources":[{"serviceAccountToken":{"expirationSeconds":3607,"path":"token"}},{"configMap":{"name":"kube-root-ca.crt","items":[{"key":"ca.crt","path":"ca.crt"}]}},{"downwardAPI":{"items":[{"path":"namespace","fieldRef":{"apiVersion":"v1","fieldPath":"metadata.namespace"}}]}},{"configMap":{"name":"openshift-service-ca.crt","items":[{"key":"service-ca.crt","path":"service-ca.crt"}]}}],"defaultMode":420}}],"containers":[{"name":"kube-rbac-proxy","image":"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69ffd8f8dcceedc2d6eb306cea33f8beabc1be1308cd5f4ee8b9a8e3eab9843","args":["--secure-listen-address=0.0.0.0:8443","--upstream=http://127.0.0.1:9090/","--tls-cert-file=/etc/tls/private/tls.crt","--tls-private-key-file=/etc/tls/private/tls.key","--logtostderr=true"],"ports":[{"name":"metrics","containerPort":8443,"protocol":"TCP"}],"resources":{"requests":{"cpu":"10m","memory":"20Mi"}},"volumeMounts":[{"name":"package-server-manager-serving-cert","mountPath":"/etc/tls/private"},{"name":"kube-api-access-ckwnn","readOnly":true,"mountPath":"/var/run/secrets/kubernetes.io/serviceaccount"}],"terminationMessagePath":"/dev/termination-log","terminationMessagePolicy":"FallbackToLogsOnError","imagePullPolicy":"IfNotPresent","securityContext":{"capabilities":{"drop":["ALL"]},"runAsUser":1000400000,"allowPrivilegeEscalation":false}},{"name":"package-server-manager","image":"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f1ca78c423f43f89a0411e40393642f64e4f8df9e5f61c25e31047c4cce170f9","command":["/bin/psm","start"],"args":["--name","$(PACKAGESERVER_NAME)","--namespace","$(PACKAGESERVER_NAMESPACE)","--metrics=:9090"],"env":[{"name":"PACKAGESERVER_NAME","value":"packageserver"},{"name":"PACKAGESERVER_IMAGE","value":"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f1ca78c423f43f89a0411e40393642f64e4f8df9e5f61c25e31047c4cce170f9"},{"name":"PACKAGESERVER_NAMESPACE","valueFrom":{"fieldRef":{"apiVersion":"v1","fieldPath":"metadata.namespace"}}},{"name":"RELEASE_VERSION","value":"4.18.29"},{"name":"GOMEMLIMIT","value":"5MiB"}],"resources":{"requests":{"cpu":"10m","memory":"10Mi"}},"volumeMounts":[{"name":"kube-api-access-ckwnn","readOnly":true,"mountPath":"/var/run/secrets/kubernetes.io/serviceaccount"}],"livenessProbe":{"httpGet":{"path":"/healthz","port":8080,"scheme":"HTTP"},"initialDelaySeconds":30,"timeoutSeconds":1,"periodSeconds":10,"successThreshold":1,"failureThreshold":3},"readinessProbe":{"httpGet":{"path":"/healthz","port":8080,"scheme":"HTTP"},"initialDelaySeconds":30,"timeoutSeconds":1,"periodSeconds":10,"successThreshold":1,"failureThreshold":3},"terminationMessagePath":"/dev/termination-log","terminationMessagePolicy":"FallbackToLogsOnError","imagePullPolicy":"IfNotPresent","securityContext":{"capabilities":{"drop":["ALL"]},"runAsUser":1000400000,"allowPrivilegeEscalation":false}}],"restartPolicy":"Always","terminationGracePeriodSeconds":30,"dnsPolicy":"ClusterFirst","nodeSelector":{"kubernetes.io/os":"linux","node-role.kubernetes.io/master":""},"serviceAccountName":"olm-operator-serviceaccount","serviceAccount":"olm-operator-serviceaccount","nodeName":"master-0","securityContext":{"seLinuxOptions":{"level":"s0:c20,c10"},"runAsNonRoot":true,"fsGroup":1000400000,"seccompProfile":{"type":"RuntimeDefault"}},"schedulerName":"default-scheduler","tolerations":[{"key":"node-role.kubernetes.io/master","operator":"Exists","effect":"NoSchedule"},{"key":"node.kubernetes.io/unreachable","operator":"Exists","effect":"NoExecute","tolerationSeconds":120},{"key":"node.kubernetes.io/not-ready","operator":"Exists","effect":"NoExecute","tolerationSeconds":120},{"key":"node.kubernetes.io/memory-pressure","operator":"Exists","effect":"NoSchedule"}],"priorityClassName":"system-cluster-critical","priority":2000000000,"enableServiceLinks":true,"preemptionPolicy":"PreemptLowerPriority"},"status":{"phase":"Running","conditions":[{"type":"PodReadyToStartContainers","status":"True","lastProbeTime":null,"lastTransitionTime":"2025-12-04T11:38:33Z"},{"type":"Initialized","status":"True","lastProbeTime":null,"lastTransitionTime":"2025-12-04T11:37:30Z"},{"type":"Ready","status":"True","lastProbeTime":null,"lastTransitionTime":"2025-12-04T11:59:30Z"},{"type":"ContainersReady","status":"True","lastProbeTime":null,"lastTransitionTime":"2025-12-04T11:59:30Z"},{"type":"PodScheduled","status":"True","lastProbeTime":null,"lastTransitionTime":"2025-12-04T11:37:30Z"}],"hostIP":"192.168.32.10","hostIPs":[{"ip":"192.168.32.10"}],"podIP":"10.128.0.13","podIPs":[{"ip":"10.128.0.13"}],"startTime":"2025-12-04T11:37:30Z","containerStatuses":[{"name":"kube-rbac-proxy","state":{"running":{"startedAt":"2025-12-04T11:38:20Z"}},"lastState":{},"ready":true,"restartCount":0,"image":"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69ffd8f8dcceedc2d6eb306cea33f8beabc1be1308cd5f4ee8b9a8e3eab9843","imageID":"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69ffd8f8dcceedc2d6eb306cea33f8beabc1be1308cd5f4ee8b9a8e3eab9843","containerID":"cri-o://f7a477bc11228fdd4daa26111c6364c8fa4e63910f723a8777b94d79118850dc","started":true,"volumeMounts":[{"name":"package-server-manager-serving-cert","mountPath":"/etc/tls/private"},{"name":"kube-api-access-ckwnn","mountPath":"/var/run/secrets/kubernetes.io/serviceaccount","readOnly":true,"recursiveReadOnly":"Disabled"}]},{"name":"package-server-manager","state":{"running":{"startedAt":"2025-12-04T11:53:54Z"}},"lastState":{"terminated":{"exitCode":1,"reason":"Error","message":"r\"}, \"highly available\": false}\n2025-12-04T11:49:30Z\tINFO\tcontrollers.packageserver\treconciliation result\t{\"csv\": {\"name\":\"packageserver\",\"namespace\":\"openshift-operator-lifecycle-manager\"}, \"res\": \"unchanged\"}\n2025-12-04T11:49:30Z\tINFO\tcontrollers.packageserver\tfinished request reconciliation\t{\"csv\": {\"name\":\"packageserver\",\"namespace\":\"openshift-operator-lifecycle-manager\"}}\nE1204 11:51:47.639922 1 leaderelection.go:429] Failed to update lock optimitically: Timeout: request did not complete within requested timeout - context deadline exceeded, falling back to slow path\nE1204 11:52:47.645428 1 leaderelection.go:436] error retrieving resource lock openshift-operator-lifecycle-manager/packageserver-controller-lock: the server was unable to return a response in the time allotted, but may still be processing the request (get leases.coordination.k8s.io packageserver-controller-lock)\nI1204 11:53:00.630259 1 leaderelection.go:297] failed to renew lease openshift-operator-lifecycle-manager/packageserver-controller-lock: timed out waiting for the condition\nE1204 11:53:34.642078 1 leaderelection.go:322] Failed to release lock: Timeout: request did not complete within requested timeout - context deadline exceeded\n2025-12-04T11:53:34Z\tINFO\tStopping and waiting for non leader election runnables\n2025-12-04T11:53:34Z\tINFO\tStopping and waiting for leader election runnables\n2025-12-04T11:53:34Z\tERROR\tsetup\tproblem running manager\t{\"error\": \"leader election lost\"}\nmain.run\n\t/build/cmd/package-server-manager/main.go:150\ngithub.com/spf13/cobra.(*Command).execute\n\t/build/vendor/github.com/spf13/cobra/command.go:985\ngithub.com/spf13/cobra.(*Command).ExecuteC\n\t/build/vendor/github.com/spf13/cobra/command.go:1117\ngithub.com/spf13/cobra.(*Command).Execute\n\t/build/vendor/github.com/spf13/cobra/command.go:1041\nmain.main\n\t/build/cmd/package-server-manager/main.go:43\nruntime.main\n\t/usr/lib/golang/src/runtime/proc.go:271\nError: leader election lost\nencountered an error while executing the binary: leader election lost\n","startedAt":"2025-12-04T11:44:28Z","finishedAt":"2025-12-04T11:53:34Z","containerID":"cri-o://d9c0a6f2319e382b7fe02ed5c5fa7e0a89c76eff9a028d94eddc90a3aa1de179"}},"ready":true,"restartCount":2,"image":"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f1ca78c423f43f89a0411e40393642f64e4f8df9e5f61c25e31047c4cce170f9","imageID":"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f1ca78c423f43f89a0411e40393642f64e4f8df9e5f61c25e31047c4cce170f9","containerID":"cri-o://b93952141b54f54e235f57ac8079e4842cc069b37f33f98481ce7921b6c9dcae","started":true,"volumeMounts":[{"name":"kube-api-access-ckwnn","mountPath":"/var/run/secrets/kubernetes.io/serviceaccount","readOnly":true,"recursiveReadOnly":"Disabled"}]}],"qosClass":"Burstable"}}olm-operator_current.log0000640000000000000000000005177415114274105033475 0ustar0000000000000000config/pod/openshift-operator-lifecycle-manager/logs/olm-operator-7cd7dbb44c-vzj4q2025-12-04T12:02:01.699873964Z time="2025-12-04T12:02:01Z" level=info msg="install strategy successful" csv=packageserver id=E/G3z namespace=openshift-operator-lifecycle-manager phase=Installing strategy=deployment 2025-12-04T12:02:01.800129337Z time="2025-12-04T12:02:01Z" level=info msg="install strategy successful" csv=packageserver id=CjewU namespace=openshift-operator-lifecycle-manager phase=Installing strategy=deployment 2025-12-04T12:02:01.899540753Z time="2025-12-04T12:02:01Z" level=info msg="install strategy successful" csv=packageserver id=07puu namespace=openshift-operator-lifecycle-manager phase=Installing strategy=deployment 2025-12-04T12:02:02.000254170Z time="2025-12-04T12:02:02Z" level=info msg="install strategy successful" csv=packageserver id=8zVGk namespace=openshift-operator-lifecycle-manager phase=Installing strategy=deployment 2025-12-04T12:02:02.170885210Z time="2025-12-04T12:02:02Z" level=info msg="install strategy successful" csv=packageserver id=D9txZ namespace=openshift-operator-lifecycle-manager phase=Installing strategy=deployment 2025-12-04T12:02:02.273041963Z time="2025-12-04T12:02:02Z" level=info msg="install strategy successful" csv=packageserver id=1xyn1 namespace=openshift-operator-lifecycle-manager phase=Installing strategy=deployment 2025-12-04T12:02:02.374913126Z time="2025-12-04T12:02:02Z" level=info msg="install strategy successful" csv=packageserver id=ILrAb namespace=openshift-operator-lifecycle-manager phase=Installing strategy=deployment 2025-12-04T12:02:02.400306220Z time="2025-12-04T12:02:02Z" level=info msg="install strategy successful" csv=packageserver id=KN9iO namespace=openshift-operator-lifecycle-manager phase=Installing strategy=deployment 2025-12-04T12:02:02.500016146Z time="2025-12-04T12:02:02Z" level=info msg="install strategy successful" csv=packageserver id=Q2NDb namespace=openshift-operator-lifecycle-manager phase=Installing strategy=deployment 2025-12-04T12:02:02.601471535Z time="2025-12-04T12:02:02Z" level=info msg="install strategy successful" csv=packageserver id=kogtc namespace=openshift-operator-lifecycle-manager phase=Installing strategy=deployment 2025-12-04T12:02:02.701035077Z time="2025-12-04T12:02:02Z" level=info msg="install strategy successful" csv=packageserver id=AVG+I namespace=openshift-operator-lifecycle-manager phase=Installing strategy=deployment 2025-12-04T12:02:02.799566925Z time="2025-12-04T12:02:02Z" level=info msg="install strategy successful" csv=packageserver id=UuZ0Z namespace=openshift-operator-lifecycle-manager phase=Installing strategy=deployment 2025-12-04T12:02:02.900589402Z time="2025-12-04T12:02:02Z" level=info msg="install strategy successful" csv=packageserver id=MhkHc namespace=openshift-operator-lifecycle-manager phase=Installing strategy=deployment 2025-12-04T12:02:03.000243555Z time="2025-12-04T12:02:03Z" level=info msg="install strategy successful" csv=packageserver id=I/l6U namespace=openshift-operator-lifecycle-manager phase=Installing strategy=deployment 2025-12-04T12:02:03.099579359Z time="2025-12-04T12:02:03Z" level=info msg="install strategy successful" csv=packageserver id=C6hdU namespace=openshift-operator-lifecycle-manager phase=Installing strategy=deployment 2025-12-04T12:02:03.198297892Z time="2025-12-04T12:02:03Z" level=info msg="install strategy successful" csv=packageserver id=FMeAt namespace=openshift-operator-lifecycle-manager phase=Installing strategy=deployment 2025-12-04T12:02:03.300073633Z time="2025-12-04T12:02:03Z" level=info msg="install strategy successful" csv=packageserver id=AXhug namespace=openshift-operator-lifecycle-manager phase=Installing strategy=deployment 2025-12-04T12:02:03.400747359Z time="2025-12-04T12:02:03Z" level=info msg="install strategy successful" csv=packageserver id=kbP9W namespace=openshift-operator-lifecycle-manager phase=Installing strategy=deployment 2025-12-04T12:02:03.499878946Z time="2025-12-04T12:02:03Z" level=info msg="install strategy successful" csv=packageserver id=nlu3x namespace=openshift-operator-lifecycle-manager phase=Installing strategy=deployment 2025-12-04T12:02:03.600085537Z time="2025-12-04T12:02:03Z" level=info msg="install strategy successful" csv=packageserver id=//LPj namespace=openshift-operator-lifecycle-manager phase=Installing strategy=deployment 2025-12-04T12:02:03.698891223Z time="2025-12-04T12:02:03Z" level=info msg="install strategy successful" csv=packageserver id=/JEnb namespace=openshift-operator-lifecycle-manager phase=Installing strategy=deployment 2025-12-04T12:02:03.800176219Z time="2025-12-04T12:02:03Z" level=info msg="install strategy successful" csv=packageserver id=EEQzw namespace=openshift-operator-lifecycle-manager phase=Installing strategy=deployment 2025-12-04T12:02:03.899977807Z time="2025-12-04T12:02:03Z" level=info msg="install strategy successful" csv=packageserver id=fOBuT namespace=openshift-operator-lifecycle-manager phase=Installing strategy=deployment 2025-12-04T12:02:03.999230378Z time="2025-12-04T12:02:03Z" level=info msg="install strategy successful" csv=packageserver id=TlOAT namespace=openshift-operator-lifecycle-manager phase=Installing strategy=deployment 2025-12-04T12:02:04.099703467Z time="2025-12-04T12:02:04Z" level=info msg="install strategy successful" csv=packageserver id=DJUKj namespace=openshift-operator-lifecycle-manager phase=Installing strategy=deployment 2025-12-04T12:02:04.199075112Z time="2025-12-04T12:02:04Z" level=info msg="install strategy successful" csv=packageserver id=Y+DlU namespace=openshift-operator-lifecycle-manager phase=Installing strategy=deployment 2025-12-04T12:02:04.298764626Z time="2025-12-04T12:02:04Z" level=info msg="install strategy successful" csv=packageserver id=BsfPg namespace=openshift-operator-lifecycle-manager phase=Installing strategy=deployment 2025-12-04T12:02:04.399755543Z time="2025-12-04T12:02:04Z" level=info msg="install strategy successful" csv=packageserver id=SXVgN namespace=openshift-operator-lifecycle-manager phase=Installing strategy=deployment 2025-12-04T12:02:04.499125978Z time="2025-12-04T12:02:04Z" level=info msg="install strategy successful" csv=packageserver id=g74wm namespace=openshift-operator-lifecycle-manager phase=Installing strategy=deployment 2025-12-04T12:02:04.598647526Z time="2025-12-04T12:02:04Z" level=info msg="install strategy successful" csv=packageserver id=LgFOx namespace=openshift-operator-lifecycle-manager phase=Installing strategy=deployment 2025-12-04T12:02:04.699107835Z time="2025-12-04T12:02:04Z" level=info msg="install strategy successful" csv=packageserver id=KQ4UG namespace=openshift-operator-lifecycle-manager phase=Installing strategy=deployment 2025-12-04T12:02:04.797951494Z time="2025-12-04T12:02:04Z" level=info msg="install strategy successful" csv=packageserver id=NgvAP namespace=openshift-operator-lifecycle-manager phase=Installing strategy=deployment 2025-12-04T12:02:04.899880239Z time="2025-12-04T12:02:04Z" level=info msg="install strategy successful" csv=packageserver id=5rgkv namespace=openshift-operator-lifecycle-manager phase=Installing strategy=deployment 2025-12-04T12:02:05.000354999Z time="2025-12-04T12:02:05Z" level=info msg="install strategy successful" csv=packageserver id=Al0FE namespace=openshift-operator-lifecycle-manager phase=Installing strategy=deployment 2025-12-04T12:02:05.099212077Z time="2025-12-04T12:02:05Z" level=info msg="install strategy successful" csv=packageserver id=whhXV namespace=openshift-operator-lifecycle-manager phase=Installing strategy=deployment 2025-12-04T12:02:05.198590562Z time="2025-12-04T12:02:05Z" level=info msg="install strategy successful" csv=packageserver id=XGTXN namespace=openshift-operator-lifecycle-manager phase=Installing strategy=deployment 2025-12-04T12:02:05.299350171Z time="2025-12-04T12:02:05Z" level=info msg="install strategy successful" csv=packageserver id=YArYo namespace=openshift-operator-lifecycle-manager phase=Installing strategy=deployment 2025-12-04T12:02:05.399724797Z time="2025-12-04T12:02:05Z" level=info msg="install strategy successful" csv=packageserver id=oBj0x namespace=openshift-operator-lifecycle-manager phase=Installing strategy=deployment 2025-12-04T12:02:05.498244954Z time="2025-12-04T12:02:05Z" level=info msg="install strategy successful" csv=packageserver id=wNfAa namespace=openshift-operator-lifecycle-manager phase=Installing strategy=deployment 2025-12-04T12:02:05.599144288Z time="2025-12-04T12:02:05Z" level=info msg="install strategy successful" csv=packageserver id=W0spi namespace=openshift-operator-lifecycle-manager phase=Installing strategy=deployment 2025-12-04T12:02:05.699468662Z time="2025-12-04T12:02:05Z" level=info msg="install strategy successful" csv=packageserver id=WP7oW namespace=openshift-operator-lifecycle-manager phase=Installing strategy=deployment 2025-12-04T12:02:05.799544519Z time="2025-12-04T12:02:05Z" level=info msg="install strategy successful" csv=packageserver id=4FHnm namespace=openshift-operator-lifecycle-manager phase=Installing strategy=deployment 2025-12-04T12:02:05.898477530Z time="2025-12-04T12:02:05Z" level=info msg="install strategy successful" csv=packageserver id=ZB/uG namespace=openshift-operator-lifecycle-manager phase=Installing strategy=deployment 2025-12-04T12:02:05.999254999Z time="2025-12-04T12:02:05Z" level=info msg="install strategy successful" csv=packageserver id=FNg5y namespace=openshift-operator-lifecycle-manager phase=Installing strategy=deployment 2025-12-04T12:02:06.098237431Z time="2025-12-04T12:02:06Z" level=info msg="install strategy successful" csv=packageserver id=nulP1 namespace=openshift-operator-lifecycle-manager phase=Installing strategy=deployment 2025-12-04T12:02:06.198746911Z time="2025-12-04T12:02:06Z" level=info msg="install strategy successful" csv=packageserver id=F+TYT namespace=openshift-operator-lifecycle-manager phase=Installing strategy=deployment 2025-12-04T12:02:06.298749676Z time="2025-12-04T12:02:06Z" level=info msg="install strategy successful" csv=packageserver id=mil/h namespace=openshift-operator-lifecycle-manager phase=Installing strategy=deployment 2025-12-04T12:02:06.399465884Z time="2025-12-04T12:02:06Z" level=info msg="install strategy successful" csv=packageserver id=pchp1 namespace=openshift-operator-lifecycle-manager phase=Installing strategy=deployment 2025-12-04T12:02:06.500074417Z time="2025-12-04T12:02:06Z" level=info msg="install strategy successful" csv=packageserver id=1Vqno namespace=openshift-operator-lifecycle-manager phase=Installing strategy=deployment 2025-12-04T12:02:06.598096680Z time="2025-12-04T12:02:06Z" level=info msg="install strategy successful" csv=packageserver id=4E0sI namespace=openshift-operator-lifecycle-manager phase=Installing strategy=deployment 2025-12-04T12:02:06.698657232Z time="2025-12-04T12:02:06Z" level=info msg="install strategy successful" csv=packageserver id=0ip8+ namespace=openshift-operator-lifecycle-manager phase=Installing strategy=deployment 2025-12-04T12:02:06.799120161Z time="2025-12-04T12:02:06Z" level=info msg="install strategy successful" csv=packageserver id=Swdi9 namespace=openshift-operator-lifecycle-manager phase=Installing strategy=deployment 2025-12-04T12:02:06.898040401Z time="2025-12-04T12:02:06Z" level=info msg="install strategy successful" csv=packageserver id=SLgLk namespace=openshift-operator-lifecycle-manager phase=Installing strategy=deployment 2025-12-04T12:02:07.001052441Z time="2025-12-04T12:02:06Z" level=info msg="install strategy successful" csv=packageserver id=kwQC0 namespace=openshift-operator-lifecycle-manager phase=Installing strategy=deployment 2025-12-04T12:02:07.098910458Z time="2025-12-04T12:02:07Z" level=info msg="install strategy successful" csv=packageserver id=9rfzO namespace=openshift-operator-lifecycle-manager phase=Installing strategy=deployment 2025-12-04T12:02:07.198440867Z time="2025-12-04T12:02:07Z" level=info msg="install strategy successful" csv=packageserver id=Lv9Tu namespace=openshift-operator-lifecycle-manager phase=Installing strategy=deployment 2025-12-04T12:02:07.299327630Z time="2025-12-04T12:02:07Z" level=info msg="install strategy successful" csv=packageserver id=dbbyB namespace=openshift-operator-lifecycle-manager phase=Installing strategy=deployment 2025-12-04T12:02:07.399618814Z time="2025-12-04T12:02:07Z" level=info msg="install strategy successful" csv=packageserver id=PBLjH namespace=openshift-operator-lifecycle-manager phase=Installing strategy=deployment 2025-12-04T12:02:07.500578818Z time="2025-12-04T12:02:07Z" level=info msg="install strategy successful" csv=packageserver id=4CdNf namespace=openshift-operator-lifecycle-manager phase=Installing strategy=deployment 2025-12-04T12:02:07.600527991Z time="2025-12-04T12:02:07Z" level=info msg="install strategy successful" csv=packageserver id=7fLMo namespace=openshift-operator-lifecycle-manager phase=Installing strategy=deployment 2025-12-04T12:02:07.699856764Z time="2025-12-04T12:02:07Z" level=info msg="install strategy successful" csv=packageserver id=2A8iv namespace=openshift-operator-lifecycle-manager phase=Installing strategy=deployment 2025-12-04T12:02:07.800166818Z time="2025-12-04T12:02:07Z" level=info msg="install strategy successful" csv=packageserver id=xu5dx namespace=openshift-operator-lifecycle-manager phase=Installing strategy=deployment 2025-12-04T12:02:07.899521373Z time="2025-12-04T12:02:07Z" level=info msg="install strategy successful" csv=packageserver id=uISFR namespace=openshift-operator-lifecycle-manager phase=Installing strategy=deployment 2025-12-04T12:02:07.999954301Z time="2025-12-04T12:02:07Z" level=info msg="install strategy successful" csv=packageserver id=zgDzH namespace=openshift-operator-lifecycle-manager phase=Installing strategy=deployment 2025-12-04T12:02:08.099710197Z time="2025-12-04T12:02:08Z" level=info msg="install strategy successful" csv=packageserver id=m6iR6 namespace=openshift-operator-lifecycle-manager phase=Installing strategy=deployment 2025-12-04T12:02:08.198587987Z time="2025-12-04T12:02:08Z" level=info msg="install strategy successful" csv=packageserver id=GO6bc namespace=openshift-operator-lifecycle-manager phase=Installing strategy=deployment 2025-12-04T12:02:08.298562290Z time="2025-12-04T12:02:08Z" level=info msg="install strategy successful" csv=packageserver id=YmW+f namespace=openshift-operator-lifecycle-manager phase=Installing strategy=deployment 2025-12-04T12:02:08.400510966Z time="2025-12-04T12:02:08Z" level=info msg="install strategy successful" csv=packageserver id=X6+r3 namespace=openshift-operator-lifecycle-manager phase=Installing strategy=deployment 2025-12-04T12:02:08.500162980Z time="2025-12-04T12:02:08Z" level=info msg="install strategy successful" csv=packageserver id=DYtpm namespace=openshift-operator-lifecycle-manager phase=Installing strategy=deployment 2025-12-04T12:02:08.599033019Z time="2025-12-04T12:02:08Z" level=info msg="install strategy successful" csv=packageserver id=CKH4Y namespace=openshift-operator-lifecycle-manager phase=Installing strategy=deployment 2025-12-04T12:02:08.700283853Z time="2025-12-04T12:02:08Z" level=info msg="install strategy successful" csv=packageserver id=NGfaC namespace=openshift-operator-lifecycle-manager phase=Installing strategy=deployment 2025-12-04T12:02:08.799304917Z time="2025-12-04T12:02:08Z" level=info msg="install strategy successful" csv=packageserver id=618M2 namespace=openshift-operator-lifecycle-manager phase=Installing strategy=deployment 2025-12-04T12:02:08.900111197Z time="2025-12-04T12:02:08Z" level=info msg="install strategy successful" csv=packageserver id=KE2EN namespace=openshift-operator-lifecycle-manager phase=Installing strategy=deployment 2025-12-04T12:02:08.999424179Z time="2025-12-04T12:02:08Z" level=info msg="install strategy successful" csv=packageserver id=IVa7a namespace=openshift-operator-lifecycle-manager phase=Installing strategy=deployment 2025-12-04T12:02:09.099107304Z time="2025-12-04T12:02:09Z" level=info msg="install strategy successful" csv=packageserver id=ExGiv namespace=openshift-operator-lifecycle-manager phase=Installing strategy=deployment 2025-12-04T12:02:09.199344566Z time="2025-12-04T12:02:09Z" level=info msg="install strategy successful" csv=packageserver id=kvJSC namespace=openshift-operator-lifecycle-manager phase=Installing strategy=deployment 2025-12-04T12:02:09.298736321Z time="2025-12-04T12:02:09Z" level=info msg="install strategy successful" csv=packageserver id=+ZQzT namespace=openshift-operator-lifecycle-manager phase=Installing strategy=deployment 2025-12-04T12:02:09.399755498Z time="2025-12-04T12:02:09Z" level=info msg="install strategy successful" csv=packageserver id=la1yS namespace=openshift-operator-lifecycle-manager phase=Installing strategy=deployment 2025-12-04T12:02:09.505136542Z time="2025-12-04T12:02:09Z" level=info msg="install strategy successful" csv=packageserver id=iEikt namespace=openshift-operator-lifecycle-manager phase=Installing strategy=deployment 2025-12-04T12:02:09.601158961Z time="2025-12-04T12:02:09Z" level=info msg="install strategy successful" csv=packageserver id=nuWNt namespace=openshift-operator-lifecycle-manager phase=Installing strategy=deployment 2025-12-04T12:02:09.699597557Z time="2025-12-04T12:02:09Z" level=info msg="install strategy successful" csv=packageserver id=SemV4 namespace=openshift-operator-lifecycle-manager phase=Installing strategy=deployment 2025-12-04T12:02:09.798384793Z time="2025-12-04T12:02:09Z" level=info msg="install strategy successful" csv=packageserver id=irhtd namespace=openshift-operator-lifecycle-manager phase=Installing strategy=deployment 2025-12-04T12:02:09.899851253Z time="2025-12-04T12:02:09Z" level=info msg="install strategy successful" csv=packageserver id=LDNI/ namespace=openshift-operator-lifecycle-manager phase=Installing strategy=deployment 2025-12-04T12:02:09.998757774Z time="2025-12-04T12:02:09Z" level=info msg="install strategy successful" csv=packageserver id=VvneD namespace=openshift-operator-lifecycle-manager phase=Installing strategy=deployment 2025-12-04T12:02:10.099738039Z time="2025-12-04T12:02:10Z" level=info msg="install strategy successful" csv=packageserver id=lxH7h namespace=openshift-operator-lifecycle-manager phase=Installing strategy=deployment 2025-12-04T12:02:10.199827206Z time="2025-12-04T12:02:10Z" level=info msg="install strategy successful" csv=packageserver id=QJvZm namespace=openshift-operator-lifecycle-manager phase=Installing strategy=deployment 2025-12-04T12:02:10.299379977Z time="2025-12-04T12:02:10Z" level=info msg="install strategy successful" csv=packageserver id=75v7w namespace=openshift-operator-lifecycle-manager phase=Installing strategy=deployment 2025-12-04T12:02:10.400318511Z time="2025-12-04T12:02:10Z" level=info msg="install strategy successful" csv=packageserver id=MCc9s namespace=openshift-operator-lifecycle-manager phase=Installing strategy=deployment 2025-12-04T12:02:10.501236545Z time="2025-12-04T12:02:10Z" level=info msg="install strategy successful" csv=packageserver id=qGSZh namespace=openshift-operator-lifecycle-manager phase=Installing strategy=deployment 2025-12-04T12:02:10.599949048Z time="2025-12-04T12:02:10Z" level=info msg="install strategy successful" csv=packageserver id=bCM1i namespace=openshift-operator-lifecycle-manager phase=Installing strategy=deployment 2025-12-04T12:02:10.698266709Z time="2025-12-04T12:02:10Z" level=info msg="install strategy successful" csv=packageserver id=D1GG6 namespace=openshift-operator-lifecycle-manager phase=Installing strategy=deployment 2025-12-04T12:02:10.798317726Z time="2025-12-04T12:02:10Z" level=info msg="install strategy successful" csv=packageserver id=c2IsL namespace=openshift-operator-lifecycle-manager phase=Installing strategy=deployment 2025-12-04T12:02:10.899516078Z time="2025-12-04T12:02:10Z" level=info msg="install strategy successful" csv=packageserver id=zQfLV namespace=openshift-operator-lifecycle-manager phase=Installing strategy=deployment 2025-12-04T12:02:10.998670066Z time="2025-12-04T12:02:10Z" level=info msg="install strategy successful" csv=packageserver id=Pkaet namespace=openshift-operator-lifecycle-manager phase=Installing strategy=deployment 2025-12-04T12:02:11.100525869Z time="2025-12-04T12:02:11Z" level=info msg="install strategy successful" csv=packageserver id=+8/dN namespace=openshift-operator-lifecycle-manager phase=Installing strategy=deployment 2025-12-04T12:02:11.198379846Z time="2025-12-04T12:02:11Z" level=info msg="install strategy successful" csv=packageserver id=W5HUY namespace=openshift-operator-lifecycle-manager phase=Installing strategy=deployment 2025-12-04T12:02:11.299886397Z time="2025-12-04T12:02:11Z" level=info msg="install strategy successful" csv=packageserver id=48SFM namespace=openshift-operator-lifecycle-manager phase=Installing strategy=deployment 2025-12-04T12:02:11.399024525Z time="2025-12-04T12:02:11Z" level=info msg="install strategy successful" csv=packageserver id=M19Zk namespace=openshift-operator-lifecycle-manager phase=Installing strategy=deployment 2025-12-04T12:02:11.499729682Z time="2025-12-04T12:02:11Z" level=info msg="install strategy successful" csv=packageserver id=uPeRU namespace=openshift-operator-lifecycle-manager phase=Installing strategy=deployment 2025-12-04T12:02:11.600688136Z time="2025-12-04T12:02:11Z" level=info msg="install strategy successful" csv=packageserver id=tUM4n namespace=openshift-operator-lifecycle-manager phase=Installing strategy=deployment catalog-operator_current.log0000640000000000000000000005076515114274105034603 0ustar0000000000000000config/pod/openshift-operator-lifecycle-manager/logs/catalog-operator-fbc6455c4-5m8ll2025-12-04T12:00:02.552089570Z time="2025-12-04T12:00:02Z" level=info msg="checking if subscriptions need update" id=075OX namespace=openshift-node 2025-12-04T12:00:02.554049082Z time="2025-12-04T12:00:02Z" level=info msg="searching for current pods" catalogsource.name=redhat-marketplace catalogsource.namespace=openshift-marketplace id=OkCMe 2025-12-04T12:00:02.554049082Z time="2025-12-04T12:00:02Z" level=info msg="evaluating current pod" catalogsource.name=redhat-marketplace catalogsource.namespace=openshift-marketplace correctHash=true correctImages=true current-pod.name=redhat-marketplace-svhl4 current-pod.namespace=openshift-marketplace id=OkCMe 2025-12-04T12:00:02.554049082Z time="2025-12-04T12:00:02Z" level=info msg="of 1 pods matching label selector, 1 have the correct images and matching hash" catalogsource.name=redhat-marketplace catalogsource.namespace=openshift-marketplace correctHash=true correctImages=true current-pod.name=redhat-marketplace-svhl4 current-pod.namespace=openshift-marketplace id=OkCMe 2025-12-04T12:00:02.554049082Z time="2025-12-04T12:00:02Z" level=info msg="checked registry server health" catalogsource.name=redhat-marketplace catalogsource.namespace=openshift-marketplace health=true id=OkCMe 2025-12-04T12:00:02.554049082Z time="2025-12-04T12:00:02Z" level=info msg="registry state good" catalogsource.name=redhat-marketplace catalogsource.namespace=openshift-marketplace id=OkCMe 2025-12-04T12:00:02.554049082Z time="2025-12-04T12:00:02Z" level=info msg="ensuring registry server" catalogsource.name=redhat-marketplace catalogsource.namespace=openshift-marketplace id=OkCMe 2025-12-04T12:00:02.741268141Z time="2025-12-04T12:00:02Z" level=info msg="No subscriptions were found in namespace openshift-network-operator" id=cvq6y namespace=openshift-network-operator 2025-12-04T12:00:02.741268141Z time="2025-12-04T12:00:02Z" level=info msg="resolving sources" id=sKANm namespace=openshift-nutanix-infra 2025-12-04T12:00:02.741268141Z time="2025-12-04T12:00:02Z" level=info msg="checking if subscriptions need update" id=sKANm namespace=openshift-nutanix-infra 2025-12-04T12:00:02.741666974Z time="2025-12-04T12:00:02Z" level=info msg="searching for current pods" catalogsource.name=certified-operators catalogsource.namespace=openshift-marketplace id=paZqW 2025-12-04T12:00:02.741910211Z time="2025-12-04T12:00:02Z" level=info msg="evaluating current pod" catalogsource.name=certified-operators catalogsource.namespace=openshift-marketplace correctHash=true correctImages=true current-pod.name=certified-operators-8qs8v current-pod.namespace=openshift-marketplace id=paZqW 2025-12-04T12:00:02.741910211Z time="2025-12-04T12:00:02Z" level=info msg="of 1 pods matching label selector, 1 have the correct images and matching hash" catalogsource.name=certified-operators catalogsource.namespace=openshift-marketplace correctHash=true correctImages=true current-pod.name=certified-operators-8qs8v current-pod.namespace=openshift-marketplace id=paZqW 2025-12-04T12:00:02.742058796Z time="2025-12-04T12:00:02Z" level=info msg="ensured registry server" catalogsource.name=certified-operators catalogsource.namespace=openshift-marketplace id=paZqW 2025-12-04T12:00:02.742058796Z time="2025-12-04T12:00:02Z" level=info msg="requeuing registry server sync based on polling interval 10m0s" catalogsource.name=certified-operators catalogsource.namespace=openshift-marketplace id=paZqW 2025-12-04T12:00:02.742139249Z time="2025-12-04T12:00:02Z" level=info msg="syncing catalog source" catalogsource.name=redhat-operators catalogsource.namespace=openshift-marketplace id=yFxVr 2025-12-04T12:00:02.742148999Z time="2025-12-04T12:00:02Z" level=info msg="synchronizing registry server" catalogsource.name=redhat-operators catalogsource.namespace=openshift-marketplace id=yFxVr 2025-12-04T12:00:03.166271701Z time="2025-12-04T12:00:03Z" level=info msg="No subscriptions were found in namespace openshift-node" id=075OX namespace=openshift-node 2025-12-04T12:00:03.166271701Z time="2025-12-04T12:00:03Z" level=info msg="resolving sources" id=9hAOW namespace=openshift-oauth-apiserver 2025-12-04T12:00:03.166312072Z time="2025-12-04T12:00:03Z" level=info msg="checking if subscriptions need update" id=9hAOW namespace=openshift-oauth-apiserver 2025-12-04T12:00:03.470307291Z time="2025-12-04T12:00:03Z" level=info msg="No subscriptions were found in namespace openshift-nutanix-infra" id=sKANm namespace=openshift-nutanix-infra 2025-12-04T12:00:03.470307291Z time="2025-12-04T12:00:03Z" level=info msg="resolving sources" id=swB24 namespace=openshift-openstack-infra 2025-12-04T12:00:03.470307291Z time="2025-12-04T12:00:03Z" level=info msg="checking if subscriptions need update" id=swB24 namespace=openshift-openstack-infra 2025-12-04T12:00:03.833924637Z time="2025-12-04T12:00:03Z" level=info msg="searching for current pods" catalogsource.name=redhat-operators catalogsource.namespace=openshift-marketplace id=yFxVr 2025-12-04T12:00:03.833924637Z time="2025-12-04T12:00:03Z" level=info msg="evaluating current pod" catalogsource.name=redhat-operators catalogsource.namespace=openshift-marketplace correctHash=true correctImages=true current-pod.name=redhat-operators-wksdw current-pod.namespace=openshift-marketplace id=yFxVr 2025-12-04T12:00:03.833924637Z time="2025-12-04T12:00:03Z" level=info msg="of 1 pods matching label selector, 1 have the correct images and matching hash" catalogsource.name=redhat-operators catalogsource.namespace=openshift-marketplace correctHash=true correctImages=true current-pod.name=redhat-operators-wksdw current-pod.namespace=openshift-marketplace id=yFxVr 2025-12-04T12:00:03.833924637Z time="2025-12-04T12:00:03Z" level=info msg="checked registry server health" catalogsource.name=redhat-operators catalogsource.namespace=openshift-marketplace health=true id=yFxVr 2025-12-04T12:00:03.833924637Z time="2025-12-04T12:00:03Z" level=info msg="registry state good" catalogsource.name=redhat-operators catalogsource.namespace=openshift-marketplace id=yFxVr 2025-12-04T12:00:03.833924637Z time="2025-12-04T12:00:03Z" level=info msg="ensuring registry server" catalogsource.name=redhat-operators catalogsource.namespace=openshift-marketplace id=yFxVr 2025-12-04T12:00:03.840564577Z time="2025-12-04T12:00:03Z" level=info msg="No subscriptions were found in namespace openshift-oauth-apiserver" id=9hAOW namespace=openshift-oauth-apiserver 2025-12-04T12:00:03.840564577Z time="2025-12-04T12:00:03Z" level=info msg="No subscriptions were found in namespace openshift-openstack-infra" id=swB24 namespace=openshift-openstack-infra 2025-12-04T12:00:03.840564577Z time="2025-12-04T12:00:03Z" level=info msg="resolving sources" id=BsOX7 namespace=openshift-operator-controller 2025-12-04T12:00:03.840564577Z time="2025-12-04T12:00:03Z" level=info msg="checking if subscriptions need update" id=BsOX7 namespace=openshift-operator-controller 2025-12-04T12:00:03.840564577Z time="2025-12-04T12:00:03Z" level=info msg="resolving sources" id=GAU6+ namespace=openshift-operator-lifecycle-manager 2025-12-04T12:00:03.840564577Z time="2025-12-04T12:00:03Z" level=info msg="checking if subscriptions need update" id=GAU6+ namespace=openshift-operator-lifecycle-manager 2025-12-04T12:00:03.843115308Z time="2025-12-04T12:00:03Z" level=info msg="No subscriptions were found in namespace openshift-operator-lifecycle-manager" id=GAU6+ namespace=openshift-operator-lifecycle-manager 2025-12-04T12:00:03.843115308Z time="2025-12-04T12:00:03Z" level=info msg="resolving sources" id=Slqtm namespace=openshift-operators 2025-12-04T12:00:03.843115308Z time="2025-12-04T12:00:03Z" level=info msg="checking if subscriptions need update" id=Slqtm namespace=openshift-operators 2025-12-04T12:00:03.843953264Z time="2025-12-04T12:00:03Z" level=info msg="searching for current pods" catalogsource.name=redhat-marketplace catalogsource.namespace=openshift-marketplace id=OkCMe 2025-12-04T12:00:03.843953264Z time="2025-12-04T12:00:03Z" level=info msg="evaluating current pod" catalogsource.name=redhat-marketplace catalogsource.namespace=openshift-marketplace correctHash=true correctImages=true current-pod.name=redhat-marketplace-svhl4 current-pod.namespace=openshift-marketplace id=OkCMe 2025-12-04T12:00:03.843953264Z time="2025-12-04T12:00:03Z" level=info msg="of 1 pods matching label selector, 1 have the correct images and matching hash" catalogsource.name=redhat-marketplace catalogsource.namespace=openshift-marketplace correctHash=true correctImages=true current-pod.name=redhat-marketplace-svhl4 current-pod.namespace=openshift-marketplace id=OkCMe 2025-12-04T12:00:03.843953264Z time="2025-12-04T12:00:03Z" level=info msg="ensured registry server" catalogsource.name=redhat-marketplace catalogsource.namespace=openshift-marketplace id=OkCMe 2025-12-04T12:00:03.843953264Z time="2025-12-04T12:00:03Z" level=info msg="requeuing registry server sync based on polling interval 10m0s" catalogsource.name=redhat-marketplace catalogsource.namespace=openshift-marketplace id=OkCMe 2025-12-04T12:00:03.843953264Z time="2025-12-04T12:00:03Z" level=info msg="syncing catalog source" catalogsource.name=community-operators catalogsource.namespace=openshift-marketplace id=DTyxI 2025-12-04T12:00:03.843998745Z time="2025-12-04T12:00:03Z" level=info msg="synchronizing registry server" catalogsource.name=community-operators catalogsource.namespace=openshift-marketplace id=DTyxI 2025-12-04T12:00:03.844613325Z time="2025-12-04T12:00:03Z" level=info msg="No subscriptions were found in namespace openshift-operator-controller" id=BsOX7 namespace=openshift-operator-controller 2025-12-04T12:00:03.844672207Z time="2025-12-04T12:00:03Z" level=info msg="resolving sources" id=fnz1u namespace=openshift-ovirt-infra 2025-12-04T12:00:03.844672207Z time="2025-12-04T12:00:03Z" level=info msg="checking if subscriptions need update" id=fnz1u namespace=openshift-ovirt-infra 2025-12-04T12:00:04.228598876Z time="2025-12-04T12:00:04Z" level=info msg="No subscriptions were found in namespace openshift-operators" id=Slqtm namespace=openshift-operators 2025-12-04T12:00:04.228598876Z time="2025-12-04T12:00:04Z" level=info msg="resolving sources" id=mbpoA namespace=openshift-ovn-kubernetes 2025-12-04T12:00:04.228598876Z time="2025-12-04T12:00:04Z" level=info msg="checking if subscriptions need update" id=mbpoA namespace=openshift-ovn-kubernetes 2025-12-04T12:00:04.228598876Z time="2025-12-04T12:00:04Z" level=info msg="No subscriptions were found in namespace openshift-ovirt-infra" id=fnz1u namespace=openshift-ovirt-infra 2025-12-04T12:00:04.228598876Z time="2025-12-04T12:00:04Z" level=info msg="resolving sources" id=0E+By namespace=openshift-route-controller-manager 2025-12-04T12:00:04.228598876Z time="2025-12-04T12:00:04Z" level=info msg="checking if subscriptions need update" id=0E+By namespace=openshift-route-controller-manager 2025-12-04T12:00:04.340884963Z time="2025-12-04T12:00:04Z" level=info msg="No subscriptions were found in namespace openshift-ovn-kubernetes" id=mbpoA namespace=openshift-ovn-kubernetes 2025-12-04T12:00:04.340884963Z time="2025-12-04T12:00:04Z" level=info msg="resolving sources" id=X3fbr namespace=openshift-service-ca 2025-12-04T12:00:04.340884963Z time="2025-12-04T12:00:04Z" level=info msg="checking if subscriptions need update" id=X3fbr namespace=openshift-service-ca 2025-12-04T12:00:04.485832343Z time="2025-12-04T12:00:04Z" level=info msg="searching for current pods" catalogsource.name=community-operators catalogsource.namespace=openshift-marketplace id=DTyxI 2025-12-04T12:00:04.485832343Z time="2025-12-04T12:00:04Z" level=info msg="evaluating current pod" catalogsource.name=community-operators catalogsource.namespace=openshift-marketplace correctHash=true correctImages=true current-pod.name=community-operators-rxhpq current-pod.namespace=openshift-marketplace id=DTyxI 2025-12-04T12:00:04.485832343Z time="2025-12-04T12:00:04Z" level=info msg="of 1 pods matching label selector, 1 have the correct images and matching hash" catalogsource.name=community-operators catalogsource.namespace=openshift-marketplace correctHash=true correctImages=true current-pod.name=community-operators-rxhpq current-pod.namespace=openshift-marketplace id=DTyxI 2025-12-04T12:00:04.485832343Z time="2025-12-04T12:00:04Z" level=info msg="checked registry server health" catalogsource.name=community-operators catalogsource.namespace=openshift-marketplace health=true id=DTyxI 2025-12-04T12:00:04.485832343Z time="2025-12-04T12:00:04Z" level=info msg="registry state good" catalogsource.name=community-operators catalogsource.namespace=openshift-marketplace id=DTyxI 2025-12-04T12:00:04.485832343Z time="2025-12-04T12:00:04Z" level=info msg="ensuring registry server" catalogsource.name=community-operators catalogsource.namespace=openshift-marketplace id=DTyxI 2025-12-04T12:00:04.525233461Z time="2025-12-04T12:00:04Z" level=info msg="No subscriptions were found in namespace openshift-route-controller-manager" id=0E+By namespace=openshift-route-controller-manager 2025-12-04T12:00:04.525233461Z time="2025-12-04T12:00:04Z" level=info msg="resolving sources" id=As6K1 namespace=openshift-service-ca-operator 2025-12-04T12:00:04.525233461Z time="2025-12-04T12:00:04Z" level=info msg="checking if subscriptions need update" id=As6K1 namespace=openshift-service-ca-operator 2025-12-04T12:00:04.661011781Z time="2025-12-04T12:00:04Z" level=info msg="searching for current pods" catalogsource.name=redhat-operators catalogsource.namespace=openshift-marketplace id=yFxVr 2025-12-04T12:00:04.661011781Z time="2025-12-04T12:00:04Z" level=info msg="evaluating current pod" catalogsource.name=redhat-operators catalogsource.namespace=openshift-marketplace correctHash=true correctImages=true current-pod.name=redhat-operators-wksdw current-pod.namespace=openshift-marketplace id=yFxVr 2025-12-04T12:00:04.661011781Z time="2025-12-04T12:00:04Z" level=info msg="of 1 pods matching label selector, 1 have the correct images and matching hash" catalogsource.name=redhat-operators catalogsource.namespace=openshift-marketplace correctHash=true correctImages=true current-pod.name=redhat-operators-wksdw current-pod.namespace=openshift-marketplace id=yFxVr 2025-12-04T12:00:04.661011781Z time="2025-12-04T12:00:04Z" level=info msg="ensured registry server" catalogsource.name=redhat-operators catalogsource.namespace=openshift-marketplace id=yFxVr 2025-12-04T12:00:04.661011781Z time="2025-12-04T12:00:04Z" level=info msg="requeuing registry server sync based on polling interval 10m0s" catalogsource.name=redhat-operators catalogsource.namespace=openshift-marketplace id=yFxVr 2025-12-04T12:00:04.661011781Z time="2025-12-04T12:00:04Z" level=info msg="syncing catalog source" catalogsource.name=certified-operators catalogsource.namespace=openshift-marketplace id=vKsTD 2025-12-04T12:00:04.661011781Z time="2025-12-04T12:00:04Z" level=info msg="synchronizing registry server" catalogsource.name=certified-operators catalogsource.namespace=openshift-marketplace id=vKsTD 2025-12-04T12:00:04.800488179Z time="2025-12-04T12:00:04Z" level=info msg="No subscriptions were found in namespace openshift-service-ca" id=X3fbr namespace=openshift-service-ca 2025-12-04T12:00:04.800488179Z time="2025-12-04T12:00:04Z" level=info msg="resolving sources" id=yW6hi namespace=openshift-user-workload-monitoring 2025-12-04T12:00:04.800488179Z time="2025-12-04T12:00:04Z" level=info msg="checking if subscriptions need update" id=yW6hi namespace=openshift-user-workload-monitoring 2025-12-04T12:00:04.941652479Z time="2025-12-04T12:00:04Z" level=info msg="No subscriptions were found in namespace openshift-service-ca-operator" id=As6K1 namespace=openshift-service-ca-operator 2025-12-04T12:00:04.941698260Z time="2025-12-04T12:00:04Z" level=info msg="resolving sources" id=BjW36 namespace=openshift-vsphere-infra 2025-12-04T12:00:04.941698260Z time="2025-12-04T12:00:04Z" level=info msg="checking if subscriptions need update" id=BjW36 namespace=openshift-vsphere-infra 2025-12-04T12:00:05.410522689Z time="2025-12-04T12:00:05Z" level=info msg="No subscriptions were found in namespace openshift-user-workload-monitoring" id=yW6hi namespace=openshift-user-workload-monitoring 2025-12-04T12:00:05.475920769Z time="2025-12-04T12:00:05Z" level=info msg="searching for current pods" catalogsource.name=certified-operators catalogsource.namespace=openshift-marketplace id=vKsTD 2025-12-04T12:00:05.476053844Z time="2025-12-04T12:00:05Z" level=info msg="evaluating current pod" catalogsource.name=certified-operators catalogsource.namespace=openshift-marketplace correctHash=true correctImages=true current-pod.name=certified-operators-8qs8v current-pod.namespace=openshift-marketplace id=vKsTD 2025-12-04T12:00:05.476053844Z time="2025-12-04T12:00:05Z" level=info msg="of 1 pods matching label selector, 1 have the correct images and matching hash" catalogsource.name=certified-operators catalogsource.namespace=openshift-marketplace correctHash=true correctImages=true current-pod.name=certified-operators-8qs8v current-pod.namespace=openshift-marketplace id=vKsTD 2025-12-04T12:00:05.476076074Z time="2025-12-04T12:00:05Z" level=info msg="checked registry server health" catalogsource.name=certified-operators catalogsource.namespace=openshift-marketplace health=true id=vKsTD 2025-12-04T12:00:05.476076074Z time="2025-12-04T12:00:05Z" level=info msg="registry state good" catalogsource.name=certified-operators catalogsource.namespace=openshift-marketplace id=vKsTD 2025-12-04T12:00:05.476085545Z time="2025-12-04T12:00:05Z" level=info msg="ensuring registry server" catalogsource.name=certified-operators catalogsource.namespace=openshift-marketplace id=vKsTD 2025-12-04T12:00:05.613613860Z time="2025-12-04T12:00:05Z" level=info msg="searching for current pods" catalogsource.name=community-operators catalogsource.namespace=openshift-marketplace id=DTyxI 2025-12-04T12:00:05.613791716Z time="2025-12-04T12:00:05Z" level=info msg="evaluating current pod" catalogsource.name=community-operators catalogsource.namespace=openshift-marketplace correctHash=true correctImages=true current-pod.name=community-operators-rxhpq current-pod.namespace=openshift-marketplace id=DTyxI 2025-12-04T12:00:05.613791716Z time="2025-12-04T12:00:05Z" level=info msg="of 1 pods matching label selector, 1 have the correct images and matching hash" catalogsource.name=community-operators catalogsource.namespace=openshift-marketplace correctHash=true correctImages=true current-pod.name=community-operators-rxhpq current-pod.namespace=openshift-marketplace id=DTyxI 2025-12-04T12:00:05.614544810Z time="2025-12-04T12:00:05Z" level=info msg="No subscriptions were found in namespace openshift-vsphere-infra" id=BjW36 namespace=openshift-vsphere-infra 2025-12-04T12:00:05.614919842Z time="2025-12-04T12:00:05Z" level=info msg="ensured registry server" catalogsource.name=community-operators catalogsource.namespace=openshift-marketplace id=DTyxI 2025-12-04T12:00:05.614919842Z time="2025-12-04T12:00:05Z" level=info msg="requeuing registry server sync based on polling interval 10m0s" catalogsource.name=community-operators catalogsource.namespace=openshift-marketplace id=DTyxI 2025-12-04T12:00:06.481838658Z time="2025-12-04T12:00:06Z" level=info msg="searching for current pods" catalogsource.name=certified-operators catalogsource.namespace=openshift-marketplace id=vKsTD 2025-12-04T12:00:06.482022184Z time="2025-12-04T12:00:06Z" level=info msg="evaluating current pod" catalogsource.name=certified-operators catalogsource.namespace=openshift-marketplace correctHash=true correctImages=true current-pod.name=certified-operators-8qs8v current-pod.namespace=openshift-marketplace id=vKsTD 2025-12-04T12:00:06.482022184Z time="2025-12-04T12:00:06Z" level=info msg="of 1 pods matching label selector, 1 have the correct images and matching hash" catalogsource.name=certified-operators catalogsource.namespace=openshift-marketplace correctHash=true correctImages=true current-pod.name=certified-operators-8qs8v current-pod.namespace=openshift-marketplace id=vKsTD 2025-12-04T12:00:06.482187670Z time="2025-12-04T12:00:06Z" level=info msg="ensured registry server" catalogsource.name=certified-operators catalogsource.namespace=openshift-marketplace id=vKsTD 2025-12-04T12:00:06.482187670Z time="2025-12-04T12:00:06Z" level=info msg="requeuing registry server sync based on polling interval 10m0s" catalogsource.name=certified-operators catalogsource.namespace=openshift-marketplace id=vKsTD 2025-12-04T12:02:02.043173448Z time="2025-12-04T12:02:02Z" level=info msg="resolving sources" id=I5wcw namespace=openshift-network-console 2025-12-04T12:02:02.043173448Z time="2025-12-04T12:02:02Z" level=info msg="checking if subscriptions need update" id=I5wcw namespace=openshift-network-console 2025-12-04T12:02:02.049920851Z time="2025-12-04T12:02:02Z" level=info msg="No subscriptions were found in namespace openshift-network-console" id=I5wcw namespace=openshift-network-console route-controller-manager_current.log0000640000000000000000000003621315114274105037422 0ustar0000000000000000config/pod/openshift-route-controller-manager/logs/route-controller-manager-95cb5f987-46bsk2025-12-04T12:00:59.170139780Z I1204 12:00:59.169933 1 cmd.go:240] Using service-serving-cert provided certificates 2025-12-04T12:00:59.170139780Z I1204 12:00:59.170098 1 leaderelection.go:121] The leader election gives 4 retries and allows for 30s of clock skew. The kube-apiserver downtime tolerance is 78s. Worst non-graceful lease acquisition is 2m43s. Worst graceful lease acquisition is {26s}. 2025-12-04T12:00:59.170983827Z I1204 12:00:59.170940 1 observer_polling.go:159] Starting file observer 2025-12-04T12:00:59.171302738Z I1204 12:00:59.171243 1 builder.go:298] route-controller-manager version 4.18.0-202511181540.p2.gce6ffa2.assembly.stream.el9-ce6ffa2-ce6ffa24dac3a5f49f70818e1159e40eaf27a6a3 2025-12-04T12:00:59.172158684Z I1204 12:00:59.172119 1 dynamic_serving_content.go:113] "Loaded a new cert/key pair" name="serving-cert::/var/run/secrets/serving-cert/tls.crt::/var/run/secrets/serving-cert/tls.key" 2025-12-04T12:00:59.861972542Z I1204 12:00:59.861909 1 requestheader_controller.go:244] Loaded a new request header values for RequestHeaderAuthRequestController 2025-12-04T12:00:59.866328969Z I1204 12:00:59.866283 1 maxinflight.go:139] "Initialized nonMutatingChan" len=400 2025-12-04T12:00:59.866393551Z I1204 12:00:59.866382 1 maxinflight.go:145] "Initialized mutatingChan" len=200 2025-12-04T12:00:59.866439583Z I1204 12:00:59.866430 1 maxinflight.go:116] "Set denominator for readonly requests" limit=400 2025-12-04T12:00:59.866463343Z I1204 12:00:59.866455 1 maxinflight.go:120] "Set denominator for mutating requests" limit=200 2025-12-04T12:00:59.870232193Z I1204 12:00:59.870186 1 secure_serving.go:57] Forcing use of http/1.1 only 2025-12-04T12:00:59.870232193Z W1204 12:00:59.870223 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected. 2025-12-04T12:00:59.870232193Z W1204 12:00:59.870227 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected. 2025-12-04T12:00:59.870257794Z W1204 12:00:59.870231 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected. 2025-12-04T12:00:59.870257794Z W1204 12:00:59.870237 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected. 2025-12-04T12:00:59.870257794Z W1204 12:00:59.870241 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected. 2025-12-04T12:00:59.870257794Z W1204 12:00:59.870243 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected. 2025-12-04T12:00:59.870283205Z I1204 12:00:59.870205 1 genericapiserver.go:514] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete 2025-12-04T12:00:59.874858809Z I1204 12:00:59.874766 1 builder.go:439] detected SingleReplicaTopologyMode, the original leader election has been altered for the default SingleReplicaTopology 2025-12-04T12:00:59.874978913Z I1204 12:00:59.874960 1 requestheader_controller.go:169] Starting RequestHeaderAuthRequestController 2025-12-04T12:00:59.875016424Z I1204 12:00:59.875004 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController 2025-12-04T12:00:59.875106967Z I1204 12:00:59.875088 1 configmap_cafile_content.go:202] "Starting controller" name="client-ca::kube-system::extension-apiserver-authentication::client-ca-file" 2025-12-04T12:00:59.875136148Z I1204 12:00:59.875126 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file 2025-12-04T12:00:59.875231141Z I1204 12:00:59.875175 1 configmap_cafile_content.go:202] "Starting controller" name="client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file" 2025-12-04T12:00:59.875263622Z I1204 12:00:59.875231 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file 2025-12-04T12:00:59.875263622Z I1204 12:00:59.875132 1 leaderelection.go:250] attempting to acquire leader lease openshift-route-controller-manager/openshift-route-controllers... 2025-12-04T12:00:59.875343455Z I1204 12:00:59.875311 1 tlsconfig.go:200] "Loaded serving cert" certName="serving-cert::/var/run/secrets/serving-cert/tls.crt::/var/run/secrets/serving-cert/tls.key" certDetail="\"route-controller-manager.openshift-route-controller-manager.svc\" [serving] validServingFor=[route-controller-manager.openshift-route-controller-manager.svc,route-controller-manager.openshift-route-controller-manager.svc.cluster.local] issuer=\"openshift-service-serving-signer@1764848266\" (2025-12-04 11:38:04 +0000 UTC to 2027-12-04 11:38:05 +0000 UTC (now=2025-12-04 12:00:59.875268872 +0000 UTC))" 2025-12-04T12:00:59.875493610Z I1204 12:00:59.875458 1 named_certificates.go:53] "Loaded SNI cert" index=0 certName="self-signed loopback" certDetail="\"apiserver-loopback-client@1764849659\" [serving] validServingFor=[apiserver-loopback-client] issuer=\"apiserver-loopback-client-ca@1764849659\" (2025-12-04 11:00:59 +0000 UTC to 2026-12-04 11:00:59 +0000 UTC (now=2025-12-04 12:00:59.875433038 +0000 UTC))" 2025-12-04T12:00:59.875590893Z I1204 12:00:59.875544 1 dynamic_serving_content.go:132] "Starting controller" name="serving-cert::/var/run/secrets/serving-cert/tls.crt::/var/run/secrets/serving-cert/tls.key" 2025-12-04T12:00:59.878333060Z I1204 12:00:59.878283 1 secure_serving.go:213] Serving securely on [::]:8443 2025-12-04T12:00:59.878350760Z I1204 12:00:59.878341 1 genericapiserver.go:662] [graceful-termination] waiting for shutdown to be initiated 2025-12-04T12:00:59.878384461Z I1204 12:00:59.878365 1 tlsconfig.go:240] "Starting DynamicServingCertificateController" 2025-12-04T12:00:59.879307391Z I1204 12:00:59.879283 1 reflector.go:359] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.30.1/tools/cache/reflector.go:232 2025-12-04T12:00:59.883743891Z I1204 12:00:59.881591 1 reflector.go:359] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.30.1/tools/cache/reflector.go:232 2025-12-04T12:00:59.883743891Z I1204 12:00:59.882743 1 reflector.go:359] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.30.1/tools/cache/reflector.go:232 2025-12-04T12:00:59.890595878Z I1204 12:00:59.890533 1 leaderelection.go:260] successfully acquired lease openshift-route-controller-manager/openshift-route-controllers 2025-12-04T12:00:59.890737043Z I1204 12:00:59.890668 1 event.go:377] Event(v1.ObjectReference{Kind:"Lease", Namespace:"openshift-route-controller-manager", Name:"openshift-route-controllers", UID:"22fec684-18bb-4301-9450-788f553f5a36", APIVersion:"coordination.k8s.io/v1", ResourceVersion:"14183", FieldPath:""}): type: 'Normal' reason: 'LeaderElection' route-controller-manager-95cb5f987-46bsk_397195f7-0562-4c91-89d0-2f6233218cb1 became leader 2025-12-04T12:00:59.892079305Z I1204 12:00:59.892050 1 controller_manager.go:36] Starting "openshift.io/ingress-ip" 2025-12-04T12:00:59.892079305Z I1204 12:00:59.892065 1 controller_manager.go:46] Started "openshift.io/ingress-ip" 2025-12-04T12:00:59.892079305Z I1204 12:00:59.892072 1 controller_manager.go:36] Starting "openshift.io/ingress-to-route" 2025-12-04T12:00:59.908626119Z I1204 12:00:59.908368 1 ingress.go:262] ingress-to-route metrics registered with prometheus 2025-12-04T12:00:59.908671250Z I1204 12:00:59.908615 1 controller_manager.go:46] Started "openshift.io/ingress-to-route" 2025-12-04T12:00:59.908671250Z I1204 12:00:59.908629 1 controller_manager.go:48] Started Route Controllers 2025-12-04T12:00:59.908995081Z I1204 12:00:59.908955 1 ingress.go:313] Starting controller 2025-12-04T12:00:59.912630396Z I1204 12:00:59.912537 1 reflector.go:359] Caches populated for *v1.IngressClass from k8s.io/client-go@v0.30.1/tools/cache/reflector.go:232 2025-12-04T12:00:59.913125962Z I1204 12:00:59.913079 1 reflector.go:359] Caches populated for *v1.Ingress from k8s.io/client-go@v0.30.1/tools/cache/reflector.go:232 2025-12-04T12:00:59.919559025Z I1204 12:00:59.919448 1 reflector.go:359] Caches populated for *v1.Service from k8s.io/client-go@v0.30.1/tools/cache/reflector.go:232 2025-12-04T12:00:59.920275178Z I1204 12:00:59.920229 1 reflector.go:359] Caches populated for *v1.Route from k8s.io/client-go@v0.30.1/tools/cache/reflector.go:232 2025-12-04T12:00:59.975345272Z I1204 12:00:59.975272 1 shared_informer.go:320] Caches are synced for RequestHeaderAuthRequestController 2025-12-04T12:00:59.975452555Z I1204 12:00:59.975280 1 shared_informer.go:320] Caches are synced for client-ca::kube-system::extension-apiserver-authentication::client-ca-file 2025-12-04T12:00:59.975602480Z I1204 12:00:59.975326 1 shared_informer.go:320] Caches are synced for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file 2025-12-04T12:00:59.976007823Z I1204 12:00:59.975888 1 tlsconfig.go:178] "Loaded client CA" index=0 certName="client-ca::kube-system::extension-apiserver-authentication::client-ca-file,client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file" certDetail="\"aggregator-signer\" [] issuer=\"\" (2025-12-04 11:26:52 +0000 UTC to 2025-12-05 11:26:52 +0000 UTC (now=2025-12-04 12:00:59.975840068 +0000 UTC))" 2025-12-04T12:00:59.976147627Z I1204 12:00:59.976115 1 tlsconfig.go:200] "Loaded serving cert" certName="serving-cert::/var/run/secrets/serving-cert/tls.crt::/var/run/secrets/serving-cert/tls.key" certDetail="\"route-controller-manager.openshift-route-controller-manager.svc\" [serving] validServingFor=[route-controller-manager.openshift-route-controller-manager.svc,route-controller-manager.openshift-route-controller-manager.svc.cluster.local] issuer=\"openshift-service-serving-signer@1764848266\" (2025-12-04 11:38:04 +0000 UTC to 2027-12-04 11:38:05 +0000 UTC (now=2025-12-04 12:00:59.976089586 +0000 UTC))" 2025-12-04T12:00:59.976291692Z I1204 12:00:59.976264 1 named_certificates.go:53] "Loaded SNI cert" index=0 certName="self-signed loopback" certDetail="\"apiserver-loopback-client@1764849659\" [serving] validServingFor=[apiserver-loopback-client] issuer=\"apiserver-loopback-client-ca@1764849659\" (2025-12-04 11:00:59 +0000 UTC to 2026-12-04 11:00:59 +0000 UTC (now=2025-12-04 12:00:59.976247141 +0000 UTC))" 2025-12-04T12:00:59.976474328Z I1204 12:00:59.976442 1 tlsconfig.go:178] "Loaded client CA" index=0 certName="client-ca::kube-system::extension-apiserver-authentication::client-ca-file,client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file" certDetail="\"admin-kubeconfig-signer\" [] issuer=\"\" (2025-12-04 11:20:50 +0000 UTC to 2035-12-02 11:20:50 +0000 UTC (now=2025-12-04 12:00:59.976424396 +0000 UTC))" 2025-12-04T12:00:59.976474328Z I1204 12:00:59.976468 1 tlsconfig.go:178] "Loaded client CA" index=1 certName="client-ca::kube-system::extension-apiserver-authentication::client-ca-file,client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file" certDetail="\"kubelet-signer\" [] issuer=\"\" (2025-12-04 11:26:53 +0000 UTC to 2025-12-05 11:26:53 +0000 UTC (now=2025-12-04 12:00:59.976456467 +0000 UTC))" 2025-12-04T12:00:59.976489228Z I1204 12:00:59.976483 1 tlsconfig.go:178] "Loaded client CA" index=2 certName="client-ca::kube-system::extension-apiserver-authentication::client-ca-file,client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file" certDetail="\"kube-control-plane-signer\" [] issuer=\"\" (2025-12-04 11:26:54 +0000 UTC to 2026-12-04 11:26:54 +0000 UTC (now=2025-12-04 12:00:59.976475178 +0000 UTC))" 2025-12-04T12:00:59.976512259Z I1204 12:00:59.976496 1 tlsconfig.go:178] "Loaded client CA" index=3 certName="client-ca::kube-system::extension-apiserver-authentication::client-ca-file,client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file" certDetail="\"kube-apiserver-to-kubelet-signer\" [] issuer=\"\" (2025-12-04 11:26:54 +0000 UTC to 2026-12-04 11:26:54 +0000 UTC (now=2025-12-04 12:00:59.976488508 +0000 UTC))" 2025-12-04T12:00:59.976512259Z I1204 12:00:59.976508 1 tlsconfig.go:178] "Loaded client CA" index=4 certName="client-ca::kube-system::extension-apiserver-authentication::client-ca-file,client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file" certDetail="\"kubelet-bootstrap-kubeconfig-signer\" [] issuer=\"\" (2025-12-04 11:26:52 +0000 UTC to 2035-12-02 11:26:52 +0000 UTC (now=2025-12-04 12:00:59.976500459 +0000 UTC))" 2025-12-04T12:00:59.976538990Z I1204 12:00:59.976523 1 tlsconfig.go:178] "Loaded client CA" index=5 certName="client-ca::kube-system::extension-apiserver-authentication::client-ca-file,client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file" certDetail="\"kube-csr-signer_@1764848267\" [] issuer=\"kubelet-signer\" (2025-12-04 11:37:46 +0000 UTC to 2025-12-05 11:26:53 +0000 UTC (now=2025-12-04 12:00:59.976513589 +0000 UTC))" 2025-12-04T12:00:59.976547980Z I1204 12:00:59.976537 1 tlsconfig.go:178] "Loaded client CA" index=6 certName="client-ca::kube-system::extension-apiserver-authentication::client-ca-file,client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file" certDetail="\"openshift-kube-apiserver-operator_node-system-admin-signer@1764848255\" [] issuer=\"\" (2025-12-04 11:37:34 +0000 UTC to 2026-12-04 11:37:35 +0000 UTC (now=2025-12-04 12:00:59.97652798 +0000 UTC))" 2025-12-04T12:00:59.976555980Z I1204 12:00:59.976550 1 tlsconfig.go:178] "Loaded client CA" index=7 certName="client-ca::kube-system::extension-apiserver-authentication::client-ca-file,client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file" certDetail="\"aggregator-signer\" [] issuer=\"\" (2025-12-04 11:26:52 +0000 UTC to 2025-12-05 11:26:52 +0000 UTC (now=2025-12-04 12:00:59.97654275 +0000 UTC))" 2025-12-04T12:00:59.976675984Z I1204 12:00:59.976656 1 tlsconfig.go:200] "Loaded serving cert" certName="serving-cert::/var/run/secrets/serving-cert/tls.crt::/var/run/secrets/serving-cert/tls.key" certDetail="\"route-controller-manager.openshift-route-controller-manager.svc\" [serving] validServingFor=[route-controller-manager.openshift-route-controller-manager.svc,route-controller-manager.openshift-route-controller-manager.svc.cluster.local] issuer=\"openshift-service-serving-signer@1764848266\" (2025-12-04 11:38:04 +0000 UTC to 2027-12-04 11:38:05 +0000 UTC (now=2025-12-04 12:00:59.976644103 +0000 UTC))" 2025-12-04T12:00:59.976764017Z I1204 12:00:59.976747 1 named_certificates.go:53] "Loaded SNI cert" index=0 certName="self-signed loopback" certDetail="\"apiserver-loopback-client@1764849659\" [serving] validServingFor=[apiserver-loopback-client] issuer=\"apiserver-loopback-client-ca@1764849659\" (2025-12-04 11:00:59 +0000 UTC to 2026-12-04 11:00:59 +0000 UTC (now=2025-12-04 12:00:59.976737546 +0000 UTC))" 2025-12-04T12:00:59.999887940Z I1204 12:00:59.999771 1 reflector.go:359] Caches populated for *v1.Secret from k8s.io/client-go@v0.30.1/tools/cache/reflector.go:232 route-controller-manager_previous.log0000640000000000000000000005726315114274105037624 0ustar0000000000000000config/pod/openshift-route-controller-manager/logs/route-controller-manager-95cb5f987-46bsk2025-12-04T11:56:43.734844443Z W1204 11:56:43.734833 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected. 2025-12-04T11:56:43.734859703Z W1204 11:56:43.734841 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected. 2025-12-04T11:56:43.734859703Z W1204 11:56:43.734847 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected. 2025-12-04T11:56:43.734859703Z W1204 11:56:43.734851 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected. 2025-12-04T11:56:43.737791532Z I1204 11:56:43.737727 1 builder.go:439] detected SingleReplicaTopologyMode, the original leader election has been altered for the default SingleReplicaTopology 2025-12-04T11:56:43.738375230Z I1204 11:56:43.738333 1 leaderelection.go:250] attempting to acquire leader lease openshift-route-controller-manager/openshift-route-controllers... 2025-12-04T11:56:43.739087371Z I1204 11:56:43.739040 1 requestheader_controller.go:169] Starting RequestHeaderAuthRequestController 2025-12-04T11:56:43.739087371Z I1204 11:56:43.739064 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController 2025-12-04T11:56:43.739109412Z I1204 11:56:43.739076 1 configmap_cafile_content.go:202] "Starting controller" name="client-ca::kube-system::extension-apiserver-authentication::client-ca-file" 2025-12-04T11:56:43.739109412Z I1204 11:56:43.739095 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file 2025-12-04T11:56:43.739160643Z I1204 11:56:43.739132 1 configmap_cafile_content.go:202] "Starting controller" name="client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file" 2025-12-04T11:56:43.739225976Z I1204 11:56:43.739171 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file 2025-12-04T11:56:43.739539815Z I1204 11:56:43.739499 1 dynamic_serving_content.go:132] "Starting controller" name="serving-cert::/var/run/secrets/serving-cert/tls.crt::/var/run/secrets/serving-cert/tls.key" 2025-12-04T11:56:43.739621158Z I1204 11:56:43.739581 1 tlsconfig.go:200] "Loaded serving cert" certName="serving-cert::/var/run/secrets/serving-cert/tls.crt::/var/run/secrets/serving-cert/tls.key" certDetail="\"route-controller-manager.openshift-route-controller-manager.svc\" [serving] validServingFor=[route-controller-manager.openshift-route-controller-manager.svc,route-controller-manager.openshift-route-controller-manager.svc.cluster.local] issuer=\"openshift-service-serving-signer@1764848266\" (2025-12-04 11:38:04 +0000 UTC to 2027-12-04 11:38:05 +0000 UTC (now=2025-12-04 11:56:43.739542645 +0000 UTC))" 2025-12-04T11:56:43.739793883Z I1204 11:56:43.739765 1 named_certificates.go:53] "Loaded SNI cert" index=0 certName="self-signed loopback" certDetail="\"apiserver-loopback-client@1764849403\" [serving] validServingFor=[apiserver-loopback-client] issuer=\"apiserver-loopback-client-ca@1764849403\" (2025-12-04 10:56:42 +0000 UTC to 2026-12-04 10:56:42 +0000 UTC (now=2025-12-04 11:56:43.739737431 +0000 UTC))" 2025-12-04T11:56:43.739828404Z I1204 11:56:43.739793 1 secure_serving.go:213] Serving securely on [::]:8443 2025-12-04T11:56:43.739842474Z I1204 11:56:43.739830 1 genericapiserver.go:662] [graceful-termination] waiting for shutdown to be initiated 2025-12-04T11:56:43.739853885Z I1204 11:56:43.739845 1 tlsconfig.go:240] "Starting DynamicServingCertificateController" 2025-12-04T11:56:43.741669951Z I1204 11:56:43.741621 1 reflector.go:359] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.30.1/tools/cache/reflector.go:232 2025-12-04T11:56:43.741711201Z I1204 11:56:43.741688 1 reflector.go:359] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.30.1/tools/cache/reflector.go:232 2025-12-04T11:56:43.741766793Z I1204 11:56:43.741710 1 reflector.go:359] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.30.1/tools/cache/reflector.go:232 2025-12-04T11:56:43.746992332Z I1204 11:56:43.746927 1 leaderelection.go:260] successfully acquired lease openshift-route-controller-manager/openshift-route-controllers 2025-12-04T11:56:43.747283091Z I1204 11:56:43.747131 1 event.go:377] Event(v1.ObjectReference{Kind:"Lease", Namespace:"openshift-route-controller-manager", Name:"openshift-route-controllers", UID:"22fec684-18bb-4301-9450-788f553f5a36", APIVersion:"coordination.k8s.io/v1", ResourceVersion:"11369", FieldPath:""}): type: 'Normal' reason: 'LeaderElection' route-controller-manager-95cb5f987-46bsk_d647a0da-6f37-43a4-80dd-2a0a58f4cd2c became leader 2025-12-04T11:56:43.748035003Z I1204 11:56:43.747990 1 controller_manager.go:36] Starting "openshift.io/ingress-ip" 2025-12-04T11:56:43.748035003Z I1204 11:56:43.748011 1 controller_manager.go:46] Started "openshift.io/ingress-ip" 2025-12-04T11:56:43.748035003Z I1204 11:56:43.748016 1 controller_manager.go:36] Starting "openshift.io/ingress-to-route" 2025-12-04T11:56:43.750666613Z I1204 11:56:43.750613 1 ingress.go:262] ingress-to-route metrics registered with prometheus 2025-12-04T11:56:43.750666613Z I1204 11:56:43.750630 1 controller_manager.go:46] Started "openshift.io/ingress-to-route" 2025-12-04T11:56:43.750666613Z I1204 11:56:43.750635 1 controller_manager.go:48] Started Route Controllers 2025-12-04T11:56:43.751094476Z I1204 11:56:43.751053 1 ingress.go:313] Starting controller 2025-12-04T11:56:43.753471798Z I1204 11:56:43.753415 1 reflector.go:359] Caches populated for *v1.IngressClass from k8s.io/client-go@v0.30.1/tools/cache/reflector.go:232 2025-12-04T11:56:43.753606972Z W1204 11:56:43.753575 1 reflector.go:547] k8s.io/client-go@v0.30.1/tools/cache/reflector.go:232: failed to list *v1.Route: the server is currently unable to handle the request (get routes.route.openshift.io) 2025-12-04T11:56:43.753617543Z E1204 11:56:43.753604 1 reflector.go:150] k8s.io/client-go@v0.30.1/tools/cache/reflector.go:232: Failed to watch *v1.Route: failed to list *v1.Route: the server is currently unable to handle the request (get routes.route.openshift.io) 2025-12-04T11:56:43.753720686Z I1204 11:56:43.753672 1 reflector.go:359] Caches populated for *v1.Ingress from k8s.io/client-go@v0.30.1/tools/cache/reflector.go:232 2025-12-04T11:56:43.764208434Z I1204 11:56:43.764104 1 reflector.go:359] Caches populated for *v1.Service from k8s.io/client-go@v0.30.1/tools/cache/reflector.go:232 2025-12-04T11:56:43.836942283Z I1204 11:56:43.836857 1 reflector.go:359] Caches populated for *v1.Secret from k8s.io/client-go@v0.30.1/tools/cache/reflector.go:232 2025-12-04T11:56:43.839207402Z I1204 11:56:43.839153 1 shared_informer.go:320] Caches are synced for client-ca::kube-system::extension-apiserver-authentication::client-ca-file 2025-12-04T11:56:43.839288275Z I1204 11:56:43.839232 1 shared_informer.go:320] Caches are synced for RequestHeaderAuthRequestController 2025-12-04T11:56:43.839321656Z I1204 11:56:43.839287 1 shared_informer.go:320] Caches are synced for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file 2025-12-04T11:56:43.839791009Z I1204 11:56:43.839754 1 tlsconfig.go:178] "Loaded client CA" index=0 certName="client-ca::kube-system::extension-apiserver-authentication::client-ca-file,client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file" certDetail="\"admin-kubeconfig-signer\" [] issuer=\"\" (2025-12-04 11:20:50 +0000 UTC to 2035-12-02 11:20:50 +0000 UTC (now=2025-12-04 11:56:43.839723187 +0000 UTC))" 2025-12-04T11:56:43.839791009Z I1204 11:56:43.839782 1 tlsconfig.go:178] "Loaded client CA" index=1 certName="client-ca::kube-system::extension-apiserver-authentication::client-ca-file,client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file" certDetail="\"kubelet-signer\" [] issuer=\"\" (2025-12-04 11:26:53 +0000 UTC to 2025-12-05 11:26:53 +0000 UTC (now=2025-12-04 11:56:43.839770999 +0000 UTC))" 2025-12-04T11:56:43.839817790Z I1204 11:56:43.839795 1 tlsconfig.go:178] "Loaded client CA" index=2 certName="client-ca::kube-system::extension-apiserver-authentication::client-ca-file,client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file" certDetail="\"kube-control-plane-signer\" [] issuer=\"\" (2025-12-04 11:26:54 +0000 UTC to 2026-12-04 11:26:54 +0000 UTC (now=2025-12-04 11:56:43.839787029 +0000 UTC))" 2025-12-04T11:56:43.839861702Z I1204 11:56:43.839835 1 tlsconfig.go:178] "Loaded client CA" index=3 certName="client-ca::kube-system::extension-apiserver-authentication::client-ca-file,client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file" certDetail="\"kube-apiserver-to-kubelet-signer\" [] issuer=\"\" (2025-12-04 11:26:54 +0000 UTC to 2026-12-04 11:26:54 +0000 UTC (now=2025-12-04 11:56:43.83980117 +0000 UTC))" 2025-12-04T11:56:43.839869892Z I1204 11:56:43.839860 1 tlsconfig.go:178] "Loaded client CA" index=4 certName="client-ca::kube-system::extension-apiserver-authentication::client-ca-file,client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file" certDetail="\"kubelet-bootstrap-kubeconfig-signer\" [] issuer=\"\" (2025-12-04 11:26:52 +0000 UTC to 2035-12-02 11:26:52 +0000 UTC (now=2025-12-04 11:56:43.839851761 +0000 UTC))" 2025-12-04T11:56:43.839894893Z I1204 11:56:43.839876 1 tlsconfig.go:178] "Loaded client CA" index=5 certName="client-ca::kube-system::extension-apiserver-authentication::client-ca-file,client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file" certDetail="\"aggregator-signer\" [] issuer=\"\" (2025-12-04 11:26:52 +0000 UTC to 2025-12-05 11:26:52 +0000 UTC (now=2025-12-04 11:56:43.839868002 +0000 UTC))" 2025-12-04T11:56:43.840062908Z I1204 11:56:43.840038 1 tlsconfig.go:200] "Loaded serving cert" certName="serving-cert::/var/run/secrets/serving-cert/tls.crt::/var/run/secrets/serving-cert/tls.key" certDetail="\"route-controller-manager.openshift-route-controller-manager.svc\" [serving] validServingFor=[route-controller-manager.openshift-route-controller-manager.svc,route-controller-manager.openshift-route-controller-manager.svc.cluster.local] issuer=\"openshift-service-serving-signer@1764848266\" (2025-12-04 11:38:04 +0000 UTC to 2027-12-04 11:38:05 +0000 UTC (now=2025-12-04 11:56:43.840022976 +0000 UTC))" 2025-12-04T11:56:43.840179711Z I1204 11:56:43.840155 1 named_certificates.go:53] "Loaded SNI cert" index=0 certName="self-signed loopback" certDetail="\"apiserver-loopback-client@1764849403\" [serving] validServingFor=[apiserver-loopback-client] issuer=\"apiserver-loopback-client-ca@1764849403\" (2025-12-04 10:56:42 +0000 UTC to 2026-12-04 10:56:42 +0000 UTC (now=2025-12-04 11:56:43.84014199 +0000 UTC))" 2025-12-04T11:56:44.837865980Z W1204 11:56:44.837762 1 reflector.go:547] k8s.io/client-go@v0.30.1/tools/cache/reflector.go:232: failed to list *v1.Route: the server is currently unable to handle the request (get routes.route.openshift.io) 2025-12-04T11:56:44.837865980Z E1204 11:56:44.837846 1 reflector.go:150] k8s.io/client-go@v0.30.1/tools/cache/reflector.go:232: Failed to watch *v1.Route: failed to list *v1.Route: the server is currently unable to handle the request (get routes.route.openshift.io) 2025-12-04T11:56:47.342496860Z W1204 11:56:47.342373 1 reflector.go:547] k8s.io/client-go@v0.30.1/tools/cache/reflector.go:232: failed to list *v1.Route: the server is currently unable to handle the request (get routes.route.openshift.io) 2025-12-04T11:56:47.342496860Z E1204 11:56:47.342429 1 reflector.go:150] k8s.io/client-go@v0.30.1/tools/cache/reflector.go:232: Failed to watch *v1.Route: failed to list *v1.Route: the server is currently unable to handle the request (get routes.route.openshift.io) 2025-12-04T11:56:51.256694188Z W1204 11:56:51.256616 1 reflector.go:547] k8s.io/client-go@v0.30.1/tools/cache/reflector.go:232: failed to list *v1.Route: the server is currently unable to handle the request (get routes.route.openshift.io) 2025-12-04T11:56:51.256694188Z E1204 11:56:51.256658 1 reflector.go:150] k8s.io/client-go@v0.30.1/tools/cache/reflector.go:232: Failed to watch *v1.Route: failed to list *v1.Route: the server is currently unable to handle the request (get routes.route.openshift.io) 2025-12-04T11:57:01.067834477Z W1204 11:57:01.067724 1 reflector.go:547] k8s.io/client-go@v0.30.1/tools/cache/reflector.go:232: failed to list *v1.Route: the server is currently unable to handle the request (get routes.route.openshift.io) 2025-12-04T11:57:01.067834477Z E1204 11:57:01.067773 1 reflector.go:150] k8s.io/client-go@v0.30.1/tools/cache/reflector.go:232: Failed to watch *v1.Route: failed to list *v1.Route: the server is currently unable to handle the request (get routes.route.openshift.io) 2025-12-04T11:57:16.039027855Z W1204 11:57:16.038927 1 reflector.go:547] k8s.io/client-go@v0.30.1/tools/cache/reflector.go:232: failed to list *v1.Route: the server is currently unable to handle the request (get routes.route.openshift.io) 2025-12-04T11:57:16.039027855Z E1204 11:57:16.038989 1 reflector.go:150] k8s.io/client-go@v0.30.1/tools/cache/reflector.go:232: Failed to watch *v1.Route: failed to list *v1.Route: the server is currently unable to handle the request (get routes.route.openshift.io) 2025-12-04T11:58:03.872350928Z W1204 11:58:03.872281 1 reflector.go:547] k8s.io/client-go@v0.30.1/tools/cache/reflector.go:232: failed to list *v1.Route: the server is currently unable to handle the request (get routes.route.openshift.io) 2025-12-04T11:58:03.872475462Z E1204 11:58:03.872460 1 reflector.go:150] k8s.io/client-go@v0.30.1/tools/cache/reflector.go:232: Failed to watch *v1.Route: failed to list *v1.Route: the server is currently unable to handle the request (get routes.route.openshift.io) 2025-12-04T11:58:48.447584053Z W1204 11:58:48.447504 1 reflector.go:547] k8s.io/client-go@v0.30.1/tools/cache/reflector.go:232: failed to list *v1.Route: the server is currently unable to handle the request (get routes.route.openshift.io) 2025-12-04T11:58:48.447584053Z E1204 11:58:48.447563 1 reflector.go:150] k8s.io/client-go@v0.30.1/tools/cache/reflector.go:232: Failed to watch *v1.Route: failed to list *v1.Route: the server is currently unable to handle the request (get routes.route.openshift.io) 2025-12-04T11:59:37.239555454Z I1204 11:59:37.239469 1 reflector.go:359] Caches populated for *v1.Route from k8s.io/client-go@v0.30.1/tools/cache/reflector.go:232 2025-12-04T11:59:49.064150484Z I1204 11:59:49.063101 1 reflector.go:359] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.30.1/tools/cache/reflector.go:232 2025-12-04T11:59:50.137854878Z I1204 11:59:50.137768 1 reflector.go:359] Caches populated for *v1.Secret from k8s.io/client-go@v0.30.1/tools/cache/reflector.go:232 2025-12-04T11:59:52.781249918Z I1204 11:59:52.781176 1 reflector.go:359] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.30.1/tools/cache/reflector.go:232 2025-12-04T11:59:52.782135176Z I1204 11:59:52.782066 1 tlsconfig.go:178] "Loaded client CA" index=0 certName="client-ca::kube-system::extension-apiserver-authentication::client-ca-file,client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file" certDetail="\"admin-kubeconfig-signer\" [] issuer=\"\" (2025-12-04 11:20:50 +0000 UTC to 2035-12-02 11:20:50 +0000 UTC (now=2025-12-04 11:59:52.782005762 +0000 UTC))" 2025-12-04T11:59:52.782154617Z I1204 11:59:52.782129 1 tlsconfig.go:178] "Loaded client CA" index=1 certName="client-ca::kube-system::extension-apiserver-authentication::client-ca-file,client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file" certDetail="\"kubelet-signer\" [] issuer=\"\" (2025-12-04 11:26:53 +0000 UTC to 2025-12-05 11:26:53 +0000 UTC (now=2025-12-04 11:59:52.782106175 +0000 UTC))" 2025-12-04T11:59:52.782188508Z I1204 11:59:52.782160 1 tlsconfig.go:178] "Loaded client CA" index=2 certName="client-ca::kube-system::extension-apiserver-authentication::client-ca-file,client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file" certDetail="\"kube-control-plane-signer\" [] issuer=\"\" (2025-12-04 11:26:54 +0000 UTC to 2026-12-04 11:26:54 +0000 UTC (now=2025-12-04 11:59:52.782139666 +0000 UTC))" 2025-12-04T11:59:52.782220999Z I1204 11:59:52.782198 1 tlsconfig.go:178] "Loaded client CA" index=3 certName="client-ca::kube-system::extension-apiserver-authentication::client-ca-file,client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file" certDetail="\"kube-apiserver-to-kubelet-signer\" [] issuer=\"\" (2025-12-04 11:26:54 +0000 UTC to 2026-12-04 11:26:54 +0000 UTC (now=2025-12-04 11:59:52.782174678 +0000 UTC))" 2025-12-04T11:59:52.782245500Z I1204 11:59:52.782233 1 tlsconfig.go:178] "Loaded client CA" index=4 certName="client-ca::kube-system::extension-apiserver-authentication::client-ca-file,client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file" certDetail="\"kubelet-bootstrap-kubeconfig-signer\" [] issuer=\"\" (2025-12-04 11:26:52 +0000 UTC to 2035-12-02 11:26:52 +0000 UTC (now=2025-12-04 11:59:52.782212339 +0000 UTC))" 2025-12-04T11:59:52.782302362Z I1204 11:59:52.782262 1 tlsconfig.go:178] "Loaded client CA" index=5 certName="client-ca::kube-system::extension-apiserver-authentication::client-ca-file,client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file" certDetail="\"kube-csr-signer_@1764848267\" [] issuer=\"kubelet-signer\" (2025-12-04 11:37:46 +0000 UTC to 2025-12-05 11:26:53 +0000 UTC (now=2025-12-04 11:59:52.78224228 +0000 UTC))" 2025-12-04T11:59:52.782313931Z I1204 11:59:52.782300 1 tlsconfig.go:178] "Loaded client CA" index=6 certName="client-ca::kube-system::extension-apiserver-authentication::client-ca-file,client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file" certDetail="\"openshift-kube-apiserver-operator_node-system-admin-signer@1764848255\" [] issuer=\"\" (2025-12-04 11:37:34 +0000 UTC to 2026-12-04 11:37:35 +0000 UTC (now=2025-12-04 11:59:52.782278701 +0000 UTC))" 2025-12-04T11:59:52.782351802Z I1204 11:59:52.782328 1 tlsconfig.go:178] "Loaded client CA" index=7 certName="client-ca::kube-system::extension-apiserver-authentication::client-ca-file,client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file" certDetail="\"aggregator-signer\" [] issuer=\"\" (2025-12-04 11:26:52 +0000 UTC to 2025-12-05 11:26:52 +0000 UTC (now=2025-12-04 11:59:52.782309872 +0000 UTC))" 2025-12-04T11:59:52.783899852Z I1204 11:59:52.782591 1 tlsconfig.go:200] "Loaded serving cert" certName="serving-cert::/var/run/secrets/serving-cert/tls.crt::/var/run/secrets/serving-cert/tls.key" certDetail="\"route-controller-manager.openshift-route-controller-manager.svc\" [serving] validServingFor=[route-controller-manager.openshift-route-controller-manager.svc,route-controller-manager.openshift-route-controller-manager.svc.cluster.local] issuer=\"openshift-service-serving-signer@1764848266\" (2025-12-04 11:38:04 +0000 UTC to 2027-12-04 11:38:05 +0000 UTC (now=2025-12-04 11:59:52.782563439 +0000 UTC))" 2025-12-04T11:59:52.783899852Z I1204 11:59:52.782795 1 named_certificates.go:53] "Loaded SNI cert" index=0 certName="self-signed loopback" certDetail="\"apiserver-loopback-client@1764849403\" [serving] validServingFor=[apiserver-loopback-client] issuer=\"apiserver-loopback-client-ca@1764849403\" (2025-12-04 10:56:42 +0000 UTC to 2026-12-04 10:56:42 +0000 UTC (now=2025-12-04 11:59:52.782770236 +0000 UTC))" 2025-12-04T11:59:53.100029983Z I1204 11:59:53.099931 1 reflector.go:359] Caches populated for *v1.Service from k8s.io/client-go@v0.30.1/tools/cache/reflector.go:232 2025-12-04T11:59:58.600617022Z I1204 11:59:58.600501 1 reflector.go:359] Caches populated for *v1.IngressClass from k8s.io/client-go@v0.30.1/tools/cache/reflector.go:232 2025-12-04T11:59:58.874357111Z I1204 11:59:58.874290 1 reflector.go:359] Caches populated for *v1.Ingress from k8s.io/client-go@v0.30.1/tools/cache/reflector.go:232 2025-12-04T12:00:00.936413550Z I1204 12:00:00.936205 1 reflector.go:359] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.30.1/tools/cache/reflector.go:232 2025-12-04T12:00:57.977344703Z I1204 12:00:57.957623 1 observer_polling.go:120] Observed file "/var/run/configmaps/config/config.yaml" has been modified (old="8946ccfb4d2c5d5a02da6c4a7b3076563b209699cfae3d98e6ef776e4419868b", new="b13d76effec9dc327c13eca786e766e224e1b685c22922cc1b02660a43d4b4c9") 2025-12-04T12:00:57.977344703Z W1204 12:00:57.957690 1 builder.go:154] Restart triggered because of file /var/run/configmaps/config/config.yaml was modified 2025-12-04T12:00:57.977344703Z I1204 12:00:57.957939 1 genericapiserver.go:670] "[graceful-termination] pre-shutdown hooks completed" name="PreShutdownHooksStopped" 2025-12-04T12:00:57.977344703Z I1204 12:00:57.958064 1 genericapiserver.go:527] "[graceful-termination] shutdown event" name="ShutdownInitiated" 2025-12-04T12:00:57.977344703Z I1204 12:00:57.958112 1 genericapiserver.go:530] "[graceful-termination] shutdown event" name="AfterShutdownDelayDuration" 2025-12-04T12:00:57.977344703Z I1204 12:00:57.958137 1 genericapiserver.go:594] "[graceful-termination] shutdown event" name="NotAcceptingNewRequest" 2025-12-04T12:00:57.977344703Z I1204 12:00:57.958147 1 genericapiserver.go:628] "[graceful-termination] not going to wait for active watch request(s) to drain" 2025-12-04T12:00:57.977344703Z I1204 12:00:57.958175 1 genericapiserver.go:619] [graceful-termination] in-flight non long-running request(s) have drained 2025-12-04T12:00:57.977344703Z I1204 12:00:57.958084 1 ingress.go:325] Shutting down controller 2025-12-04T12:00:57.977344703Z I1204 12:00:57.958193 1 genericapiserver.go:660] "[graceful-termination] shutdown event" name="InFlightRequestsDrained" 2025-12-04T12:00:57.977344703Z I1204 12:00:57.958281 1 tlsconfig.go:255] "Shutting down DynamicServingCertificateController" 2025-12-04T12:00:57.977344703Z I1204 12:00:57.958319 1 configmap_cafile_content.go:223] "Shutting down controller" name="client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file" 2025-12-04T12:00:57.977344703Z I1204 12:00:57.958336 1 secure_serving.go:258] Stopped listening on [::]:8443 2025-12-04T12:00:57.977344703Z I1204 12:00:57.958349 1 requestheader_controller.go:183] Shutting down RequestHeaderAuthRequestController 2025-12-04T12:00:57.977344703Z I1204 12:00:57.958353 1 genericapiserver.go:577] "[graceful-termination] shutdown event" name="HTTPServerStoppedListening" 2025-12-04T12:00:57.977344703Z I1204 12:00:57.958364 1 configmap_cafile_content.go:223] "Shutting down controller" name="client-ca::kube-system::extension-apiserver-authentication::client-ca-file" 2025-12-04T12:00:57.977344703Z I1204 12:00:57.958289 1 object_count_tracker.go:151] "StorageObjectCountTracker pruner is exiting" 2025-12-04T12:00:57.977344703Z I1204 12:00:57.958432 1 dynamic_serving_content.go:146] "Shutting down controller" name="serving-cert::/var/run/secrets/serving-cert/tls.crt::/var/run/secrets/serving-cert/tls.key" 2025-12-04T12:00:57.977344703Z I1204 12:00:57.958465 1 genericapiserver.go:690] [graceful-termination] apiserver is exiting 2025-12-04T12:00:57.977344703Z I1204 12:00:57.958476 1 builder.go:329] server exited 2025-12-04T12:00:57.977344703Z W1204 12:00:57.975064 1 leaderelection.go:84] leader election lost config/pod/openshift-route-controller-manager/route-controller-manager-95cb5f987-46bsk.json0000640000000000000000000001534415114274105030236 0ustar0000000000000000{"metadata":{"name":"route-controller-manager-95cb5f987-46bsk","generateName":"route-controller-manager-95cb5f987-","namespace":"openshift-route-controller-manager","uid":"e4d7939a-5961-4608-b910-73e71aa55bf6","resourceVersion":"14186","creationTimestamp":"2025-12-04T11:38:45Z","labels":{"app":"route-controller-manager","openshift-route-controller-manager-anti-affinity":"true","pod-template-hash":"95cb5f987","route-controller-manager":"true"},"annotations":{"configmaps/client-ca":"6705","configmaps/config":"9058","k8s.ovn.org/pod-networks":"{\"default\":{\"ip_addresses\":[\"10.128.0.70/23\"],\"mac_address\":\"0a:58:0a:80:00:46\",\"gateway_ips\":[\"10.128.0.1\"],\"routes\":[{\"dest\":\"10.128.0.0/16\",\"nextHop\":\"10.128.0.1\"},{\"dest\":\"172.30.0.0/16\",\"nextHop\":\"10.128.0.1\"},{\"dest\":\"169.254.0.5/32\",\"nextHop\":\"10.128.0.1\"},{\"dest\":\"100.64.0.0/16\",\"nextHop\":\"10.128.0.1\"}],\"ip_address\":\"10.128.0.70/23\",\"gateway_ip\":\"10.128.0.1\",\"role\":\"primary\"}}","k8s.v1.cni.cncf.io/network-status":"[{\n \"name\": \"ovn-kubernetes\",\n \"interface\": \"eth0\",\n \"ips\": [\n \"10.128.0.70\"\n ],\n \"mac\": \"0a:58:0a:80:00:46\",\n \"default\": true,\n \"dns\": {}\n}]","openshift.io/required-scc":"restricted-v2","openshift.io/scc":"restricted-v2","openshiftcontrollermanagers.operator.openshift.io/cluster":"2","seccomp.security.alpha.kubernetes.io/pod":"runtime/default"},"ownerReferences":[{"apiVersion":"apps/v1","kind":"ReplicaSet","name":"route-controller-manager-95cb5f987","uid":"87c53bab-9ff1-4030-a813-e3538487a655","controller":true,"blockOwnerDeletion":true}]},"spec":{"volumes":[{"name":"config","configMap":{"name":"config","defaultMode":420}},{"name":"client-ca","configMap":{"name":"client-ca","defaultMode":420}},{"name":"serving-cert","secret":{"secretName":"serving-cert","defaultMode":420}},{"name":"kube-api-access-d77m8","projected":{"sources":[{"serviceAccountToken":{"expirationSeconds":3607,"path":"token"}},{"configMap":{"name":"kube-root-ca.crt","items":[{"key":"ca.crt","path":"ca.crt"}]}},{"downwardAPI":{"items":[{"path":"namespace","fieldRef":{"apiVersion":"v1","fieldPath":"metadata.namespace"}}]}},{"configMap":{"name":"openshift-service-ca.crt","items":[{"key":"service-ca.crt","path":"service-ca.crt"}]}}],"defaultMode":420}}],"containers":[{"name":"route-controller-manager","image":"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c416b201d480bddb5a4960ec42f4740761a1335001cf84ba5ae19ad6857771b1","command":["route-controller-manager","start"],"args":["--config=/var/run/configmaps/config/config.yaml","-v=2"],"ports":[{"containerPort":8443,"protocol":"TCP"}],"env":[{"name":"POD_NAME","valueFrom":{"fieldRef":{"apiVersion":"v1","fieldPath":"metadata.name"}}}],"resources":{"requests":{"cpu":"100m","memory":"100Mi"}},"volumeMounts":[{"name":"config","mountPath":"/var/run/configmaps/config"},{"name":"client-ca","mountPath":"/var/run/configmaps/client-ca"},{"name":"serving-cert","mountPath":"/var/run/secrets/serving-cert"},{"name":"kube-api-access-d77m8","readOnly":true,"mountPath":"/var/run/secrets/kubernetes.io/serviceaccount"}],"livenessProbe":{"httpGet":{"path":"healthz","port":8443,"scheme":"HTTPS"},"initialDelaySeconds":30,"timeoutSeconds":1,"periodSeconds":10,"successThreshold":1,"failureThreshold":3},"readinessProbe":{"httpGet":{"path":"healthz","port":8443,"scheme":"HTTPS"},"timeoutSeconds":1,"periodSeconds":10,"successThreshold":1,"failureThreshold":10},"terminationMessagePath":"/dev/termination-log","terminationMessagePolicy":"FallbackToLogsOnError","imagePullPolicy":"IfNotPresent","securityContext":{"capabilities":{"drop":["ALL"]},"runAsUser":1000580000,"allowPrivilegeEscalation":false}}],"restartPolicy":"Always","terminationGracePeriodSeconds":30,"dnsPolicy":"ClusterFirst","nodeSelector":{"node-role.kubernetes.io/master":""},"serviceAccountName":"route-controller-manager-sa","serviceAccount":"route-controller-manager-sa","nodeName":"master-0","securityContext":{"seLinuxOptions":{"level":"s0:c24,c14"},"runAsNonRoot":true,"fsGroup":1000580000,"seccompProfile":{"type":"RuntimeDefault"}},"imagePullSecrets":[{"name":"route-controller-manager-sa-dockercfg-qzw27"}],"affinity":{"podAntiAffinity":{"requiredDuringSchedulingIgnoredDuringExecution":[{"labelSelector":{"matchLabels":{"app":"route-controller-manager","openshift-route-controller-manager-anti-affinity":"true","route-controller-manager":"true"}},"topologyKey":"kubernetes.io/hostname"}]}},"schedulerName":"default-scheduler","tolerations":[{"key":"node-role.kubernetes.io/master","operator":"Exists","effect":"NoSchedule"},{"key":"node.kubernetes.io/unreachable","operator":"Exists","effect":"NoExecute","tolerationSeconds":120},{"key":"node.kubernetes.io/not-ready","operator":"Exists","effect":"NoExecute","tolerationSeconds":120},{"key":"node.kubernetes.io/memory-pressure","operator":"Exists","effect":"NoSchedule"}],"priorityClassName":"system-node-critical","priority":2000001000,"enableServiceLinks":true,"preemptionPolicy":"PreemptLowerPriority"},"status":{"phase":"Running","conditions":[{"type":"PodReadyToStartContainers","status":"True","lastProbeTime":null,"lastTransitionTime":"2025-12-04T11:45:12Z"},{"type":"Initialized","status":"True","lastProbeTime":null,"lastTransitionTime":"2025-12-04T11:38:50Z"},{"type":"Ready","status":"True","lastProbeTime":null,"lastTransitionTime":"2025-12-04T12:00:59Z"},{"type":"ContainersReady","status":"True","lastProbeTime":null,"lastTransitionTime":"2025-12-04T12:00:59Z"},{"type":"PodScheduled","status":"True","lastProbeTime":null,"lastTransitionTime":"2025-12-04T11:38:50Z"}],"hostIP":"192.168.32.10","hostIPs":[{"ip":"192.168.32.10"}],"podIP":"10.128.0.70","podIPs":[{"ip":"10.128.0.70"}],"startTime":"2025-12-04T11:38:50Z","containerStatuses":[{"name":"route-controller-manager","state":{"running":{"startedAt":"2025-12-04T12:00:59Z"}},"lastState":{"terminated":{"exitCode":0,"reason":"Completed","startedAt":"2025-12-04T11:56:42Z","finishedAt":"2025-12-04T12:00:58Z","containerID":"cri-o://028cb113b8a4d0739d18a17371ebd3e1e120e22b132d20c18e57ba6391cc21d7"}},"ready":true,"restartCount":3,"image":"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c416b201d480bddb5a4960ec42f4740761a1335001cf84ba5ae19ad6857771b1","imageID":"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c416b201d480bddb5a4960ec42f4740761a1335001cf84ba5ae19ad6857771b1","containerID":"cri-o://388a36eac650c8c09011f149057b67b4396ceec6d06ae7dcaa9fd13f4018f169","started":true,"volumeMounts":[{"name":"config","mountPath":"/var/run/configmaps/config"},{"name":"client-ca","mountPath":"/var/run/configmaps/client-ca"},{"name":"serving-cert","mountPath":"/var/run/secrets/serving-cert"},{"name":"kube-api-access-d77m8","mountPath":"/var/run/secrets/kubernetes.io/serviceaccount","readOnly":true,"recursiveReadOnly":"Disabled"}]}],"qosClass":"Burstable"}}controller-manager_current.log0000640000000000000000000006030515114274105034165 0ustar0000000000000000config/pod/openshift-controller-manager/logs/controller-manager-6686654b8d-rrndk2025-12-04T11:56:05.931330254Z W1204 11:56:05.931236 1 reflector.go:561] k8s.io/client-go@v1.5.2/tools/cache/reflector.go:243: failed to list *v1.TemplateInstance: the server is currently unable to handle the request (get templateinstances.template.openshift.io) 2025-12-04T11:56:05.931330254Z E1204 11:56:05.931289 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go@v1.5.2/tools/cache/reflector.go:243: Failed to watch *v1.TemplateInstance: failed to list *v1.TemplateInstance: the server is currently unable to handle the request (get templateinstances.template.openshift.io)" 2025-12-04T11:56:06.464883157Z W1204 11:56:06.464776 1 reflector.go:561] k8s.io/client-go@v1.5.2/tools/cache/reflector.go:243: failed to list *v1.Build: the server is currently unable to handle the request (get builds.build.openshift.io) 2025-12-04T11:56:06.464944659Z E1204 11:56:06.464884 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go@v1.5.2/tools/cache/reflector.go:243: Failed to watch *v1.Build: failed to list *v1.Build: the server is currently unable to handle the request (get builds.build.openshift.io)" 2025-12-04T11:56:06.851365794Z W1204 11:56:06.851191 1 reflector.go:561] k8s.io/client-go@v1.5.2/tools/cache/reflector.go:243: failed to list *v1.BuildConfig: the server is currently unable to handle the request (get buildconfigs.build.openshift.io) 2025-12-04T11:56:06.851365794Z E1204 11:56:06.851266 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go@v1.5.2/tools/cache/reflector.go:243: Failed to watch *v1.BuildConfig: failed to list *v1.BuildConfig: the server is currently unable to handle the request (get buildconfigs.build.openshift.io)" 2025-12-04T11:56:22.546954748Z W1204 11:56:22.546769 1 reflector.go:561] k8s.io/client-go@v1.5.2/tools/cache/reflector.go:243: failed to list *v1.Image: the server is currently unable to handle the request (get images.image.openshift.io) 2025-12-04T11:56:22.546954748Z E1204 11:56:22.546848 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go@v1.5.2/tools/cache/reflector.go:243: Failed to watch *v1.Image: failed to list *v1.Image: the server is currently unable to handle the request (get images.image.openshift.io)" 2025-12-04T11:56:22.677604146Z W1204 11:56:22.677514 1 reflector.go:561] k8s.io/client-go@v1.5.2/tools/cache/reflector.go:243: failed to list *v1.ImageStream: the server is currently unable to handle the request (get imagestreams.image.openshift.io) 2025-12-04T11:56:22.677604146Z E1204 11:56:22.677579 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go@v1.5.2/tools/cache/reflector.go:243: Failed to watch *v1.ImageStream: failed to list *v1.ImageStream: the server is currently unable to handle the request (get imagestreams.image.openshift.io)" 2025-12-04T11:56:22.967763258Z W1204 11:56:22.967689 1 reflector.go:561] k8s.io/client-go@v1.5.2/tools/cache/reflector.go:243: failed to list *v1.DeploymentConfig: the server is currently unable to handle the request (get deploymentconfigs.apps.openshift.io) 2025-12-04T11:56:22.967763258Z E1204 11:56:22.967742 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go@v1.5.2/tools/cache/reflector.go:243: Failed to watch *v1.DeploymentConfig: failed to list *v1.DeploymentConfig: the server is currently unable to handle the request (get deploymentconfigs.apps.openshift.io)" 2025-12-04T11:56:23.800547128Z W1204 11:56:23.800458 1 reflector.go:561] k8s.io/client-go@v1.5.2/tools/cache/reflector.go:243: failed to list *v1.TemplateInstance: the server is currently unable to handle the request (get templateinstances.template.openshift.io) 2025-12-04T11:56:23.800547128Z E1204 11:56:23.800518 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go@v1.5.2/tools/cache/reflector.go:243: Failed to watch *v1.TemplateInstance: failed to list *v1.TemplateInstance: the server is currently unable to handle the request (get templateinstances.template.openshift.io)" 2025-12-04T11:56:28.035948080Z W1204 11:56:28.035865 1 reflector.go:561] k8s.io/client-go@v1.5.2/tools/cache/reflector.go:243: failed to list *v1.Build: the server is currently unable to handle the request (get builds.build.openshift.io) 2025-12-04T11:56:28.035948080Z E1204 11:56:28.035922 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go@v1.5.2/tools/cache/reflector.go:243: Failed to watch *v1.Build: failed to list *v1.Build: the server is currently unable to handle the request (get builds.build.openshift.io)" 2025-12-04T11:56:31.567966020Z W1204 11:56:31.567887 1 reflector.go:561] k8s.io/client-go@v1.5.2/tools/cache/reflector.go:243: failed to list *v1.BuildConfig: the server is currently unable to handle the request (get buildconfigs.build.openshift.io) 2025-12-04T11:56:31.567966020Z E1204 11:56:31.567952 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go@v1.5.2/tools/cache/reflector.go:243: Failed to watch *v1.BuildConfig: failed to list *v1.BuildConfig: the server is currently unable to handle the request (get buildconfigs.build.openshift.io)" 2025-12-04T11:56:55.035777867Z W1204 11:56:55.035682 1 reflector.go:561] k8s.io/client-go@v1.5.2/tools/cache/reflector.go:243: failed to list *v1.DeploymentConfig: the server is currently unable to handle the request (get deploymentconfigs.apps.openshift.io) 2025-12-04T11:56:55.035863450Z E1204 11:56:55.035766 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go@v1.5.2/tools/cache/reflector.go:243: Failed to watch *v1.DeploymentConfig: failed to list *v1.DeploymentConfig: the server is currently unable to handle the request (get deploymentconfigs.apps.openshift.io)" 2025-12-04T11:56:59.449159847Z W1204 11:56:59.449075 1 reflector.go:561] k8s.io/client-go@v1.5.2/tools/cache/reflector.go:243: failed to list *v1.TemplateInstance: the server is currently unable to handle the request (get templateinstances.template.openshift.io) 2025-12-04T11:56:59.449159847Z E1204 11:56:59.449139 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go@v1.5.2/tools/cache/reflector.go:243: Failed to watch *v1.TemplateInstance: failed to list *v1.TemplateInstance: the server is currently unable to handle the request (get templateinstances.template.openshift.io)" 2025-12-04T11:57:03.909205286Z W1204 11:57:03.909068 1 reflector.go:561] k8s.io/client-go@v1.5.2/tools/cache/reflector.go:243: failed to list *v1.Image: the server is currently unable to handle the request (get images.image.openshift.io) 2025-12-04T11:57:03.909205286Z E1204 11:57:03.909137 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go@v1.5.2/tools/cache/reflector.go:243: Failed to watch *v1.Image: failed to list *v1.Image: the server is currently unable to handle the request (get images.image.openshift.io)" 2025-12-04T11:57:09.375995280Z W1204 11:57:09.375893 1 reflector.go:561] k8s.io/client-go@v1.5.2/tools/cache/reflector.go:243: failed to list *v1.ImageStream: the server is currently unable to handle the request (get imagestreams.image.openshift.io) 2025-12-04T11:57:09.375995280Z E1204 11:57:09.375950 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go@v1.5.2/tools/cache/reflector.go:243: Failed to watch *v1.ImageStream: failed to list *v1.ImageStream: the server is currently unable to handle the request (get imagestreams.image.openshift.io)" 2025-12-04T11:57:10.205647383Z W1204 11:57:10.205552 1 reflector.go:561] k8s.io/client-go@v1.5.2/tools/cache/reflector.go:243: failed to list *v1.BuildConfig: the server is currently unable to handle the request (get buildconfigs.build.openshift.io) 2025-12-04T11:57:10.205647383Z E1204 11:57:10.205616 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go@v1.5.2/tools/cache/reflector.go:243: Failed to watch *v1.BuildConfig: failed to list *v1.BuildConfig: the server is currently unable to handle the request (get buildconfigs.build.openshift.io)" 2025-12-04T11:57:15.200454813Z W1204 11:57:15.200333 1 reflector.go:561] k8s.io/client-go@v1.5.2/tools/cache/reflector.go:243: failed to list *v1.Build: the server is currently unable to handle the request (get builds.build.openshift.io) 2025-12-04T11:57:15.200454813Z E1204 11:57:15.200400 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go@v1.5.2/tools/cache/reflector.go:243: Failed to watch *v1.Build: failed to list *v1.Build: the server is currently unable to handle the request (get builds.build.openshift.io)" 2025-12-04T11:57:31.761214428Z W1204 11:57:31.761154 1 reflector.go:561] k8s.io/client-go@v1.5.2/tools/cache/reflector.go:243: failed to list *v1.TemplateInstance: the server is currently unable to handle the request (get templateinstances.template.openshift.io) 2025-12-04T11:57:31.761274389Z E1204 11:57:31.761210 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go@v1.5.2/tools/cache/reflector.go:243: Failed to watch *v1.TemplateInstance: failed to list *v1.TemplateInstance: the server is currently unable to handle the request (get templateinstances.template.openshift.io)" 2025-12-04T11:57:39.554879698Z W1204 11:57:39.554778 1 reflector.go:561] k8s.io/client-go@v1.5.2/tools/cache/reflector.go:243: failed to list *v1.DeploymentConfig: the server is currently unable to handle the request (get deploymentconfigs.apps.openshift.io) 2025-12-04T11:57:39.554931900Z E1204 11:57:39.554875 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go@v1.5.2/tools/cache/reflector.go:243: Failed to watch *v1.DeploymentConfig: failed to list *v1.DeploymentConfig: the server is currently unable to handle the request (get deploymentconfigs.apps.openshift.io)" 2025-12-04T11:57:43.920607696Z W1204 11:57:43.920526 1 reflector.go:561] k8s.io/client-go@v1.5.2/tools/cache/reflector.go:243: failed to list *v1.ImageStream: the server is currently unable to handle the request (get imagestreams.image.openshift.io) 2025-12-04T11:57:43.920607696Z E1204 11:57:43.920582 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go@v1.5.2/tools/cache/reflector.go:243: Failed to watch *v1.ImageStream: failed to list *v1.ImageStream: the server is currently unable to handle the request (get imagestreams.image.openshift.io)" 2025-12-04T11:57:48.693784012Z W1204 11:57:48.693674 1 reflector.go:561] k8s.io/client-go@v1.5.2/tools/cache/reflector.go:243: failed to list *v1.Image: the server is currently unable to handle the request (get images.image.openshift.io) 2025-12-04T11:57:48.693784012Z E1204 11:57:48.693741 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go@v1.5.2/tools/cache/reflector.go:243: Failed to watch *v1.Image: failed to list *v1.Image: the server is currently unable to handle the request (get images.image.openshift.io)" 2025-12-04T11:57:50.011496499Z W1204 11:57:50.010770 1 reflector.go:561] k8s.io/client-go@v1.5.2/tools/cache/reflector.go:243: failed to list *v1.BuildConfig: the server is currently unable to handle the request (get buildconfigs.build.openshift.io) 2025-12-04T11:57:50.011609092Z E1204 11:57:50.011564 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go@v1.5.2/tools/cache/reflector.go:243: Failed to watch *v1.BuildConfig: failed to list *v1.BuildConfig: the server is currently unable to handle the request (get buildconfigs.build.openshift.io)" 2025-12-04T11:58:10.822013801Z W1204 11:58:10.821895 1 reflector.go:561] k8s.io/client-go@v1.5.2/tools/cache/reflector.go:243: failed to list *v1.Build: the server is currently unable to handle the request (get builds.build.openshift.io) 2025-12-04T11:58:10.822013801Z E1204 11:58:10.821962 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go@v1.5.2/tools/cache/reflector.go:243: Failed to watch *v1.Build: failed to list *v1.Build: the server is currently unable to handle the request (get builds.build.openshift.io)" 2025-12-04T11:58:12.740564575Z W1204 11:58:12.740479 1 reflector.go:561] k8s.io/client-go@v1.5.2/tools/cache/reflector.go:243: failed to list *v1.TemplateInstance: the server is currently unable to handle the request (get templateinstances.template.openshift.io) 2025-12-04T11:58:12.740564575Z E1204 11:58:12.740539 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go@v1.5.2/tools/cache/reflector.go:243: Failed to watch *v1.TemplateInstance: failed to list *v1.TemplateInstance: the server is currently unable to handle the request (get templateinstances.template.openshift.io)" 2025-12-04T11:58:15.481013586Z W1204 11:58:15.480336 1 reflector.go:561] k8s.io/client-go@v1.5.2/tools/cache/reflector.go:243: failed to list *v1.DeploymentConfig: the server is currently unable to handle the request (get deploymentconfigs.apps.openshift.io) 2025-12-04T11:58:15.481013586Z E1204 11:58:15.480407 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go@v1.5.2/tools/cache/reflector.go:243: Failed to watch *v1.DeploymentConfig: failed to list *v1.DeploymentConfig: the server is currently unable to handle the request (get deploymentconfigs.apps.openshift.io)" 2025-12-04T11:58:30.367652608Z W1204 11:58:30.367544 1 reflector.go:561] k8s.io/client-go@v1.5.2/tools/cache/reflector.go:243: failed to list *v1.ImageStream: the server is currently unable to handle the request (get imagestreams.image.openshift.io) 2025-12-04T11:58:30.367652608Z E1204 11:58:30.367609 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go@v1.5.2/tools/cache/reflector.go:243: Failed to watch *v1.ImageStream: failed to list *v1.ImageStream: the server is currently unable to handle the request (get imagestreams.image.openshift.io)" 2025-12-04T11:58:33.920333854Z W1204 11:58:33.920232 1 reflector.go:561] k8s.io/client-go@v1.5.2/tools/cache/reflector.go:243: failed to list *v1.Image: the server is currently unable to handle the request (get images.image.openshift.io) 2025-12-04T11:58:33.920333854Z E1204 11:58:33.920305 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go@v1.5.2/tools/cache/reflector.go:243: Failed to watch *v1.Image: failed to list *v1.Image: the server is currently unable to handle the request (get images.image.openshift.io)" 2025-12-04T11:58:42.878978418Z W1204 11:58:42.878787 1 reflector.go:561] k8s.io/client-go@v1.5.2/tools/cache/reflector.go:243: failed to list *v1.BuildConfig: the server is currently unable to handle the request (get buildconfigs.build.openshift.io) 2025-12-04T11:58:42.878978418Z E1204 11:58:42.878891 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go@v1.5.2/tools/cache/reflector.go:243: Failed to watch *v1.BuildConfig: failed to list *v1.BuildConfig: the server is currently unable to handle the request (get buildconfigs.build.openshift.io)" 2025-12-04T11:58:46.467776465Z W1204 11:58:46.467666 1 reflector.go:561] k8s.io/client-go@v1.5.2/tools/cache/reflector.go:243: failed to list *v1.Build: the server is currently unable to handle the request (get builds.build.openshift.io) 2025-12-04T11:58:46.467776465Z E1204 11:58:46.467740 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go@v1.5.2/tools/cache/reflector.go:243: Failed to watch *v1.Build: failed to list *v1.Build: the server is currently unable to handle the request (get builds.build.openshift.io)" 2025-12-04T11:59:11.196935749Z W1204 11:59:11.196861 1 reflector.go:561] k8s.io/client-go@v1.5.2/tools/cache/reflector.go:243: failed to list *v1.DeploymentConfig: the server is currently unable to handle the request (get deploymentconfigs.apps.openshift.io) 2025-12-04T11:59:11.196979370Z E1204 11:59:11.196933 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go@v1.5.2/tools/cache/reflector.go:243: Failed to watch *v1.DeploymentConfig: failed to list *v1.DeploymentConfig: the server is currently unable to handle the request (get deploymentconfigs.apps.openshift.io)" 2025-12-04T11:59:12.472496907Z W1204 11:59:12.472395 1 reflector.go:561] k8s.io/client-go@v1.5.2/tools/cache/reflector.go:243: failed to list *v1.TemplateInstance: the server is currently unable to handle the request (get templateinstances.template.openshift.io) 2025-12-04T11:59:12.472576889Z E1204 11:59:12.472484 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go@v1.5.2/tools/cache/reflector.go:243: Failed to watch *v1.TemplateInstance: failed to list *v1.TemplateInstance: the server is currently unable to handle the request (get templateinstances.template.openshift.io)" 2025-12-04T11:59:12.505112479Z W1204 11:59:12.505003 1 reflector.go:561] k8s.io/client-go@v1.5.2/tools/cache/reflector.go:243: failed to list *v1.ImageStream: the server is currently unable to handle the request (get imagestreams.image.openshift.io) 2025-12-04T11:59:12.505112479Z E1204 11:59:12.505064 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go@v1.5.2/tools/cache/reflector.go:243: Failed to watch *v1.ImageStream: failed to list *v1.ImageStream: the server is currently unable to handle the request (get imagestreams.image.openshift.io)" 2025-12-04T11:59:17.778139192Z W1204 11:59:17.777985 1 reflector.go:561] k8s.io/client-go@v1.5.2/tools/cache/reflector.go:243: failed to list *v1.BuildConfig: Get "https://172.30.0.1:443/apis/build.openshift.io/v1/buildconfigs?limit=500&resourceVersion=0": dial tcp 172.30.0.1:443: connect: connection refused 2025-12-04T11:59:17.778139192Z E1204 11:59:17.778124 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go@v1.5.2/tools/cache/reflector.go:243: Failed to watch *v1.BuildConfig: failed to list *v1.BuildConfig: Get \"https://172.30.0.1:443/apis/build.openshift.io/v1/buildconfigs?limit=500&resourceVersion=0\": dial tcp 172.30.0.1:443: connect: connection refused" 2025-12-04T11:59:20.237043657Z E1204 11:59:20.236966 1 leaderelection.go:429] Failed to update lock optimitically: Put "https://172.30.0.1:443/apis/coordination.k8s.io/v1/namespaces/openshift-controller-manager/leases/openshift-master-controllers": dial tcp 172.30.0.1:443: connect: connection refused, falling back to slow path 2025-12-04T11:59:20.237522681Z E1204 11:59:20.237488 1 leaderelection.go:436] error retrieving resource lock openshift-controller-manager/openshift-master-controllers: Get "https://172.30.0.1:443/apis/coordination.k8s.io/v1/namespaces/openshift-controller-manager/leases/openshift-master-controllers": dial tcp 172.30.0.1:443: connect: connection refused 2025-12-04T11:59:23.124248587Z W1204 11:59:23.124115 1 reflector.go:561] k8s.io/client-go@v1.5.2/tools/cache/reflector.go:243: failed to list *v1.Image: Get "https://172.30.0.1:443/apis/image.openshift.io/v1/images?limit=500&resourceVersion=0": dial tcp 172.30.0.1:443: connect: connection refused 2025-12-04T11:59:23.124248587Z E1204 11:59:23.124232 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go@v1.5.2/tools/cache/reflector.go:243: Failed to watch *v1.Image: failed to list *v1.Image: Get \"https://172.30.0.1:443/apis/image.openshift.io/v1/images?limit=500&resourceVersion=0\": dial tcp 172.30.0.1:443: connect: connection refused" 2025-12-04T11:59:42.583202875Z W1204 11:59:42.583113 1 warnings.go:70] apps.openshift.io/v1 DeploymentConfig is deprecated in v4.14+, unavailable in v4.10000+ 2025-12-04T11:59:42.583268597Z I1204 11:59:42.583236 1 reflector.go:368] Caches populated for *v1.DeploymentConfig from k8s.io/client-go@v1.5.2/tools/cache/reflector.go:243 2025-12-04T11:59:42.589901496Z W1204 11:59:42.589798 1 warnings.go:70] apps.openshift.io/v1 DeploymentConfig is deprecated in v4.14+, unavailable in v4.10000+ 2025-12-04T11:59:42.660558854Z I1204 11:59:42.660453 1 factory.go:85] deploymentconfig controller caches are synced. Starting workers. 2025-12-04T11:59:45.549329484Z I1204 11:59:45.549243 1 reflector.go:368] Caches populated for *v1.Build from k8s.io/client-go@v1.5.2/tools/cache/reflector.go:243 2025-12-04T11:59:46.202386038Z I1204 11:59:46.202287 1 reflector.go:368] Caches populated for *v1.ImageStream from k8s.io/client-go@v1.5.2/tools/cache/reflector.go:243 2025-12-04T11:59:46.301072963Z I1204 11:59:46.300966 1 build_controller.go:503] Starting build controller 2025-12-04T11:59:46.301072963Z I1204 11:59:46.301005 1 build_controller.go:505] OpenShift image registry hostname: 2025-12-04T11:59:46.314994254Z I1204 11:59:46.314901 1 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go@v1.5.2/tools/cache/reflector.go:243 2025-12-04T11:59:48.077875477Z I1204 11:59:48.077753 1 reflector.go:368] Caches populated for *v1.DaemonSet from k8s.io/client-go@v1.5.2/tools/cache/reflector.go:243 2025-12-04T11:59:48.560546974Z I1204 11:59:48.560197 1 reflector.go:368] Caches populated for *v1.ReplicationController from k8s.io/client-go@v1.5.2/tools/cache/reflector.go:243 2025-12-04T11:59:49.951249399Z I1204 11:59:49.951160 1 reflector.go:368] Caches populated for *v1.CronJob from k8s.io/client-go@v1.5.2/tools/cache/reflector.go:243 2025-12-04T11:59:50.377603371Z I1204 11:59:50.377510 1 reflector.go:368] Caches populated for *v1.BuildConfig from k8s.io/client-go@v1.5.2/tools/cache/reflector.go:243 2025-12-04T11:59:50.429323639Z I1204 11:59:50.429236 1 buildconfig_controller.go:212] Starting buildconfig controller 2025-12-04T11:59:51.055445059Z I1204 11:59:51.055331 1 reflector.go:368] Caches populated for *v1.ImageTagMirrorSet from k8s.io/client-go@v1.5.2/tools/cache/reflector.go:243 2025-12-04T11:59:52.250031204Z I1204 11:59:52.247586 1 reflector.go:368] Caches populated for *v1.Deployment from k8s.io/client-go@v1.5.2/tools/cache/reflector.go:243 2025-12-04T11:59:52.994206022Z I1204 11:59:52.994110 1 reflector.go:368] Caches populated for *v1.StatefulSet from k8s.io/client-go@v1.5.2/tools/cache/reflector.go:243 2025-12-04T11:59:54.466136869Z I1204 11:59:54.466055 1 reflector.go:368] Caches populated for *v1.ImageDigestMirrorSet from k8s.io/client-go@v1.5.2/tools/cache/reflector.go:243 2025-12-04T11:59:55.543762938Z I1204 11:59:55.543664 1 reflector.go:368] Caches populated for *v1.Pod from k8s.io/client-go@v1.5.2/tools/cache/reflector.go:243 2025-12-04T11:59:56.231134799Z I1204 11:59:56.231022 1 reflector.go:368] Caches populated for *v1.Secret from k8s.io/client-go@v1.5.2/tools/cache/reflector.go:243 2025-12-04T11:59:56.342427823Z I1204 11:59:56.342327 1 reflector.go:368] Caches populated for *v1alpha1.ImageContentSourcePolicy from k8s.io/client-go@v1.5.2/tools/cache/reflector.go:243 2025-12-04T11:59:56.399074467Z I1204 11:59:56.398991 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v1.5.2/tools/cache/reflector.go:243 2025-12-04T11:59:57.385850299Z I1204 11:59:57.385717 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v1.5.2/tools/cache/reflector.go:243 2025-12-04T11:59:57.805830962Z I1204 11:59:57.671620 1 reflector.go:368] Caches populated for *v1.Build from k8s.io/client-go@v1.5.2/tools/cache/reflector.go:243 2025-12-04T11:59:57.810402347Z I1204 11:59:57.810321 1 reflector.go:368] Caches populated for *v1.Proxy from k8s.io/client-go@v1.5.2/tools/cache/reflector.go:243 2025-12-04T11:59:58.670760094Z I1204 11:59:58.670657 1 reflector.go:368] Caches populated for *v1.Namespace from k8s.io/client-go@v1.5.2/tools/cache/reflector.go:243 2025-12-04T11:59:58.834365186Z I1204 11:59:58.834268 1 reflector.go:368] Caches populated for *v1.ServiceAccount from k8s.io/client-go@v1.5.2/tools/cache/reflector.go:243 2025-12-04T11:59:59.732859182Z I1204 11:59:59.732724 1 reflector.go:368] Caches populated for *v1.RoleBinding from k8s.io/client-go@v1.5.2/tools/cache/reflector.go:243 2025-12-04T12:00:00.287538349Z I1204 12:00:00.287447 1 reflector.go:368] Caches populated for *v1.Image from k8s.io/client-go@v1.5.2/tools/cache/reflector.go:243 2025-12-04T12:00:03.050514835Z I1204 12:00:03.050427 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v1.5.2/tools/cache/reflector.go:243 2025-12-04T12:00:08.582870931Z I1204 12:00:08.582708 1 reflector.go:368] Caches populated for *v1.Image from k8s.io/client-go@v1.5.2/tools/cache/reflector.go:243 2025-12-04T12:00:10.074287585Z I1204 12:00:10.074186 1 reflector.go:368] Caches populated for *v1.TemplateInstance from k8s.io/client-go@v1.5.2/tools/cache/reflector.go:243 2025-12-04T12:00:10.084435887Z I1204 12:00:10.084329 1 templateinstance_controller.go:297] Starting TemplateInstance controller 2025-12-04T12:00:10.167860319Z I1204 12:00:10.167684 1 templateinstance_finalizer.go:194] Starting TemplateInstanceFinalizer controller controller-manager_previous.log0000640000000000000000000006666615114274105034377 0ustar0000000000000000config/pod/openshift-controller-manager/logs/controller-manager-6686654b8d-rrndk2025-12-04T11:47:44.172931982Z E1204 11:47:44.172912 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go@v1.5.2/tools/cache/reflector.go:243: Failed to watch *v1.TemplateInstance: failed to list *v1.TemplateInstance: the server is currently unable to handle the request (get templateinstances.template.openshift.io)" 2025-12-04T11:47:57.113493991Z W1204 11:47:57.113336 1 reflector.go:561] k8s.io/client-go@v1.5.2/tools/cache/reflector.go:243: failed to list *v1.Build: the server is currently unable to handle the request (get builds.build.openshift.io) 2025-12-04T11:47:57.113493991Z E1204 11:47:57.113399 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go@v1.5.2/tools/cache/reflector.go:243: Failed to watch *v1.Build: failed to list *v1.Build: the server is currently unable to handle the request (get builds.build.openshift.io)" 2025-12-04T11:48:14.584735406Z W1204 11:48:14.584630 1 reflector.go:561] k8s.io/client-go@v1.5.2/tools/cache/reflector.go:243: failed to list *v1.ImageStream: the server is currently unable to handle the request (get imagestreams.image.openshift.io) 2025-12-04T11:48:14.584735406Z E1204 11:48:14.584711 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go@v1.5.2/tools/cache/reflector.go:243: Failed to watch *v1.ImageStream: failed to list *v1.ImageStream: the server is currently unable to handle the request (get imagestreams.image.openshift.io)" 2025-12-04T11:48:24.904984931Z W1204 11:48:24.904876 1 reflector.go:561] k8s.io/client-go@v1.5.2/tools/cache/reflector.go:243: failed to list *v1.DeploymentConfig: the server is currently unable to handle the request (get deploymentconfigs.apps.openshift.io) 2025-12-04T11:48:24.904984931Z E1204 11:48:24.904940 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go@v1.5.2/tools/cache/reflector.go:243: Failed to watch *v1.DeploymentConfig: failed to list *v1.DeploymentConfig: the server is currently unable to handle the request (get deploymentconfigs.apps.openshift.io)" 2025-12-04T11:48:27.200729143Z W1204 11:48:27.200618 1 reflector.go:561] k8s.io/client-go@v1.5.2/tools/cache/reflector.go:243: failed to list *v1.Image: the server is currently unable to handle the request (get images.image.openshift.io) 2025-12-04T11:48:27.200729143Z E1204 11:48:27.200680 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go@v1.5.2/tools/cache/reflector.go:243: Failed to watch *v1.Image: failed to list *v1.Image: the server is currently unable to handle the request (get images.image.openshift.io)" 2025-12-04T11:48:30.524705954Z W1204 11:48:30.524578 1 reflector.go:561] k8s.io/client-go@v1.5.2/tools/cache/reflector.go:243: failed to list *v1.TemplateInstance: the server is currently unable to handle the request (get templateinstances.template.openshift.io) 2025-12-04T11:48:30.524705954Z E1204 11:48:30.524637 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go@v1.5.2/tools/cache/reflector.go:243: Failed to watch *v1.TemplateInstance: failed to list *v1.TemplateInstance: the server is currently unable to handle the request (get templateinstances.template.openshift.io)" 2025-12-04T11:48:32.008419023Z W1204 11:48:32.008325 1 reflector.go:561] k8s.io/client-go@v1.5.2/tools/cache/reflector.go:243: failed to list *v1.Build: the server is currently unable to handle the request (get builds.build.openshift.io) 2025-12-04T11:48:32.008419023Z E1204 11:48:32.008391 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go@v1.5.2/tools/cache/reflector.go:243: Failed to watch *v1.Build: failed to list *v1.Build: the server is currently unable to handle the request (get builds.build.openshift.io)" 2025-12-04T11:48:37.141323160Z W1204 11:48:37.141201 1 reflector.go:561] k8s.io/client-go@v1.5.2/tools/cache/reflector.go:243: failed to list *v1.BuildConfig: the server is currently unable to handle the request (get buildconfigs.build.openshift.io) 2025-12-04T11:48:37.141323160Z E1204 11:48:37.141251 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go@v1.5.2/tools/cache/reflector.go:243: Failed to watch *v1.BuildConfig: failed to list *v1.BuildConfig: the server is currently unable to handle the request (get buildconfigs.build.openshift.io)" 2025-12-04T11:48:58.774098918Z W1204 11:48:58.774024 1 reflector.go:561] k8s.io/client-go@v1.5.2/tools/cache/reflector.go:243: failed to list *v1.ImageStream: the server is currently unable to handle the request (get imagestreams.image.openshift.io) 2025-12-04T11:48:58.774148629Z E1204 11:48:58.774100 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go@v1.5.2/tools/cache/reflector.go:243: Failed to watch *v1.ImageStream: failed to list *v1.ImageStream: the server is currently unable to handle the request (get imagestreams.image.openshift.io)" 2025-12-04T11:49:00.096557131Z W1204 11:49:00.096477 1 reflector.go:561] k8s.io/client-go@v1.5.2/tools/cache/reflector.go:243: failed to list *v1.Image: the server is currently unable to handle the request (get images.image.openshift.io) 2025-12-04T11:49:00.096557131Z E1204 11:49:00.096535 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go@v1.5.2/tools/cache/reflector.go:243: Failed to watch *v1.Image: failed to list *v1.Image: the server is currently unable to handle the request (get images.image.openshift.io)" 2025-12-04T11:49:02.829340503Z W1204 11:49:02.829169 1 reflector.go:561] k8s.io/client-go@v1.5.2/tools/cache/reflector.go:243: failed to list *v1.DeploymentConfig: the server is currently unable to handle the request (get deploymentconfigs.apps.openshift.io) 2025-12-04T11:49:02.829340503Z E1204 11:49:02.829250 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go@v1.5.2/tools/cache/reflector.go:243: Failed to watch *v1.DeploymentConfig: failed to list *v1.DeploymentConfig: the server is currently unable to handle the request (get deploymentconfigs.apps.openshift.io)" 2025-12-04T11:49:04.637075563Z W1204 11:49:04.636925 1 reflector.go:561] k8s.io/client-go@v1.5.2/tools/cache/reflector.go:243: failed to list *v1.TemplateInstance: the server is currently unable to handle the request (get templateinstances.template.openshift.io) 2025-12-04T11:49:04.637075563Z E1204 11:49:04.637025 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go@v1.5.2/tools/cache/reflector.go:243: Failed to watch *v1.TemplateInstance: failed to list *v1.TemplateInstance: the server is currently unable to handle the request (get templateinstances.template.openshift.io)" 2025-12-04T11:49:21.532129062Z W1204 11:49:21.532035 1 reflector.go:561] k8s.io/client-go@v1.5.2/tools/cache/reflector.go:243: failed to list *v1.BuildConfig: the server is currently unable to handle the request (get buildconfigs.build.openshift.io) 2025-12-04T11:49:21.532179254Z E1204 11:49:21.532131 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go@v1.5.2/tools/cache/reflector.go:243: Failed to watch *v1.BuildConfig: failed to list *v1.BuildConfig: the server is currently unable to handle the request (get buildconfigs.build.openshift.io)" 2025-12-04T11:49:27.745168502Z W1204 11:49:27.745093 1 reflector.go:561] k8s.io/client-go@v1.5.2/tools/cache/reflector.go:243: failed to list *v1.Build: the server is currently unable to handle the request (get builds.build.openshift.io) 2025-12-04T11:49:27.745209813Z E1204 11:49:27.745165 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go@v1.5.2/tools/cache/reflector.go:243: Failed to watch *v1.Build: failed to list *v1.Build: the server is currently unable to handle the request (get builds.build.openshift.io)" 2025-12-04T11:49:45.168774947Z W1204 11:49:45.168576 1 reflector.go:561] k8s.io/client-go@v1.5.2/tools/cache/reflector.go:243: failed to list *v1.Image: the server is currently unable to handle the request (get images.image.openshift.io) 2025-12-04T11:49:45.168774947Z E1204 11:49:45.168711 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go@v1.5.2/tools/cache/reflector.go:243: Failed to watch *v1.Image: failed to list *v1.Image: the server is currently unable to handle the request (get images.image.openshift.io)" 2025-12-04T11:49:47.712712780Z W1204 11:49:47.712623 1 reflector.go:561] k8s.io/client-go@v1.5.2/tools/cache/reflector.go:243: failed to list *v1.ImageStream: the server is currently unable to handle the request (get imagestreams.image.openshift.io) 2025-12-04T11:49:47.712712780Z E1204 11:49:47.712688 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go@v1.5.2/tools/cache/reflector.go:243: Failed to watch *v1.ImageStream: failed to list *v1.ImageStream: the server is currently unable to handle the request (get imagestreams.image.openshift.io)" 2025-12-04T11:49:53.789551439Z W1204 11:49:53.789450 1 reflector.go:561] k8s.io/client-go@v1.5.2/tools/cache/reflector.go:243: failed to list *v1.DeploymentConfig: the server is currently unable to handle the request (get deploymentconfigs.apps.openshift.io) 2025-12-04T11:49:53.789551439Z E1204 11:49:53.789508 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go@v1.5.2/tools/cache/reflector.go:243: Failed to watch *v1.DeploymentConfig: failed to list *v1.DeploymentConfig: the server is currently unable to handle the request (get deploymentconfigs.apps.openshift.io)" 2025-12-04T11:49:56.888569865Z W1204 11:49:56.888469 1 reflector.go:561] k8s.io/client-go@v1.5.2/tools/cache/reflector.go:243: failed to list *v1.TemplateInstance: the server is currently unable to handle the request (get templateinstances.template.openshift.io) 2025-12-04T11:49:56.888569865Z E1204 11:49:56.888528 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go@v1.5.2/tools/cache/reflector.go:243: Failed to watch *v1.TemplateInstance: failed to list *v1.TemplateInstance: the server is currently unable to handle the request (get templateinstances.template.openshift.io)" 2025-12-04T11:50:09.816149673Z W1204 11:50:09.816021 1 reflector.go:561] k8s.io/client-go@v1.5.2/tools/cache/reflector.go:243: failed to list *v1.BuildConfig: the server is currently unable to handle the request (get buildconfigs.build.openshift.io) 2025-12-04T11:50:09.816149673Z E1204 11:50:09.816085 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go@v1.5.2/tools/cache/reflector.go:243: Failed to watch *v1.BuildConfig: failed to list *v1.BuildConfig: the server is currently unable to handle the request (get buildconfigs.build.openshift.io)" 2025-12-04T11:50:15.817875402Z W1204 11:50:15.817776 1 reflector.go:561] k8s.io/client-go@v1.5.2/tools/cache/reflector.go:243: failed to list *v1.Build: the server is currently unable to handle the request (get builds.build.openshift.io) 2025-12-04T11:50:15.817936314Z E1204 11:50:15.817869 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go@v1.5.2/tools/cache/reflector.go:243: Failed to watch *v1.Build: failed to list *v1.Build: the server is currently unable to handle the request (get builds.build.openshift.io)" 2025-12-04T11:50:22.879875231Z W1204 11:50:22.879793 1 reflector.go:561] k8s.io/client-go@v1.5.2/tools/cache/reflector.go:243: failed to list *v1.Image: the server is currently unable to handle the request (get images.image.openshift.io) 2025-12-04T11:50:22.880023796Z E1204 11:50:22.880007 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go@v1.5.2/tools/cache/reflector.go:243: Failed to watch *v1.Image: failed to list *v1.Image: the server is currently unable to handle the request (get images.image.openshift.io)" 2025-12-04T11:50:28.397454771Z W1204 11:50:28.397357 1 reflector.go:561] k8s.io/client-go@v1.5.2/tools/cache/reflector.go:243: failed to list *v1.DeploymentConfig: the server is currently unable to handle the request (get deploymentconfigs.apps.openshift.io) 2025-12-04T11:50:28.397454771Z E1204 11:50:28.397420 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go@v1.5.2/tools/cache/reflector.go:243: Failed to watch *v1.DeploymentConfig: failed to list *v1.DeploymentConfig: the server is currently unable to handle the request (get deploymentconfigs.apps.openshift.io)" 2025-12-04T11:50:30.580652017Z W1204 11:50:30.580551 1 reflector.go:561] k8s.io/client-go@v1.5.2/tools/cache/reflector.go:243: failed to list *v1.TemplateInstance: the server is currently unable to handle the request (get templateinstances.template.openshift.io) 2025-12-04T11:50:30.580652017Z E1204 11:50:30.580626 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go@v1.5.2/tools/cache/reflector.go:243: Failed to watch *v1.TemplateInstance: failed to list *v1.TemplateInstance: the server is currently unable to handle the request (get templateinstances.template.openshift.io)" 2025-12-04T11:50:31.959687537Z W1204 11:50:31.959607 1 reflector.go:561] k8s.io/client-go@v1.5.2/tools/cache/reflector.go:243: failed to list *v1.ImageStream: the server is currently unable to handle the request (get imagestreams.image.openshift.io) 2025-12-04T11:50:31.959687537Z E1204 11:50:31.959667 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go@v1.5.2/tools/cache/reflector.go:243: Failed to watch *v1.ImageStream: failed to list *v1.ImageStream: the server is currently unable to handle the request (get imagestreams.image.openshift.io)" 2025-12-04T11:50:47.920199081Z W1204 11:50:47.920073 1 reflector.go:561] k8s.io/client-go@v1.5.2/tools/cache/reflector.go:243: failed to list *v1.Build: the server is currently unable to handle the request (get builds.build.openshift.io) 2025-12-04T11:50:47.920199081Z E1204 11:50:47.920157 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go@v1.5.2/tools/cache/reflector.go:243: Failed to watch *v1.Build: failed to list *v1.Build: the server is currently unable to handle the request (get builds.build.openshift.io)" 2025-12-04T11:51:01.494390988Z W1204 11:51:01.494246 1 reflector.go:561] k8s.io/client-go@v1.5.2/tools/cache/reflector.go:243: failed to list *v1.BuildConfig: the server is currently unable to handle the request (get buildconfigs.build.openshift.io) 2025-12-04T11:51:01.494390988Z E1204 11:51:01.494342 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go@v1.5.2/tools/cache/reflector.go:243: Failed to watch *v1.BuildConfig: failed to list *v1.BuildConfig: the server is currently unable to handle the request (get buildconfigs.build.openshift.io)" 2025-12-04T11:51:05.955969502Z W1204 11:51:05.955868 1 reflector.go:561] k8s.io/client-go@v1.5.2/tools/cache/reflector.go:243: failed to list *v1.ImageStream: the server is currently unable to handle the request (get imagestreams.image.openshift.io) 2025-12-04T11:51:05.955969502Z E1204 11:51:05.955926 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go@v1.5.2/tools/cache/reflector.go:243: Failed to watch *v1.ImageStream: failed to list *v1.ImageStream: the server is currently unable to handle the request (get imagestreams.image.openshift.io)" 2025-12-04T11:51:10.015702911Z W1204 11:51:10.015606 1 reflector.go:561] k8s.io/client-go@v1.5.2/tools/cache/reflector.go:243: failed to list *v1.Image: the server is currently unable to handle the request (get images.image.openshift.io) 2025-12-04T11:51:10.015702911Z E1204 11:51:10.015677 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go@v1.5.2/tools/cache/reflector.go:243: Failed to watch *v1.Image: failed to list *v1.Image: the server is currently unable to handle the request (get images.image.openshift.io)" 2025-12-04T11:51:10.780634842Z W1204 11:51:10.780556 1 reflector.go:561] k8s.io/client-go@v1.5.2/tools/cache/reflector.go:243: failed to list *v1.TemplateInstance: the server is currently unable to handle the request (get templateinstances.template.openshift.io) 2025-12-04T11:51:10.780634842Z E1204 11:51:10.780618 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go@v1.5.2/tools/cache/reflector.go:243: Failed to watch *v1.TemplateInstance: failed to list *v1.TemplateInstance: the server is currently unable to handle the request (get templateinstances.template.openshift.io)" 2025-12-04T11:51:17.013708670Z W1204 11:51:17.013619 1 reflector.go:561] k8s.io/client-go@v1.5.2/tools/cache/reflector.go:243: failed to list *v1.DeploymentConfig: the server is currently unable to handle the request (get deploymentconfigs.apps.openshift.io) 2025-12-04T11:51:17.013708670Z E1204 11:51:17.013687 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go@v1.5.2/tools/cache/reflector.go:243: Failed to watch *v1.DeploymentConfig: failed to list *v1.DeploymentConfig: the server is currently unable to handle the request (get deploymentconfigs.apps.openshift.io)" 2025-12-04T11:51:23.000680781Z W1204 11:51:23.000584 1 reflector.go:561] k8s.io/client-go@v1.5.2/tools/cache/reflector.go:243: failed to list *v1.Build: the server is currently unable to handle the request (get builds.build.openshift.io) 2025-12-04T11:51:23.000680781Z E1204 11:51:23.000644 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go@v1.5.2/tools/cache/reflector.go:243: Failed to watch *v1.Build: failed to list *v1.Build: the server is currently unable to handle the request (get builds.build.openshift.io)" 2025-12-04T11:51:49.005333929Z W1204 11:51:49.005267 1 reflector.go:561] k8s.io/client-go@v1.5.2/tools/cache/reflector.go:243: failed to list *v1.ImageStream: the server is currently unable to handle the request (get imagestreams.image.openshift.io) 2025-12-04T11:51:49.005407881Z E1204 11:51:49.005326 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go@v1.5.2/tools/cache/reflector.go:243: Failed to watch *v1.ImageStream: failed to list *v1.ImageStream: the server is currently unable to handle the request (get imagestreams.image.openshift.io)" 2025-12-04T11:51:50.427001926Z W1204 11:51:50.426899 1 reflector.go:561] k8s.io/client-go@v1.5.2/tools/cache/reflector.go:243: failed to list *v1.Image: the server is currently unable to handle the request (get images.image.openshift.io) 2025-12-04T11:51:50.427001926Z E1204 11:51:50.426978 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go@v1.5.2/tools/cache/reflector.go:243: Failed to watch *v1.Image: failed to list *v1.Image: the server is currently unable to handle the request (get images.image.openshift.io)" 2025-12-04T11:51:50.609897034Z E1204 11:51:50.609689 1 leaderelection.go:429] Failed to update lock optimitically: Timeout: request did not complete within requested timeout - context deadline exceeded, falling back to slow path 2025-12-04T11:51:50.883914003Z W1204 11:51:50.883702 1 reflector.go:561] k8s.io/client-go@v1.5.2/tools/cache/reflector.go:243: failed to list *v1.BuildConfig: the server is currently unable to handle the request (get buildconfigs.build.openshift.io) 2025-12-04T11:51:50.883914003Z E1204 11:51:50.883765 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go@v1.5.2/tools/cache/reflector.go:243: Failed to watch *v1.BuildConfig: failed to list *v1.BuildConfig: the server is currently unable to handle the request (get buildconfigs.build.openshift.io)" 2025-12-04T11:51:54.262210841Z W1204 11:51:54.262092 1 reflector.go:561] k8s.io/client-go@v1.5.2/tools/cache/reflector.go:243: failed to list *v1.Build: the server is currently unable to handle the request (get builds.build.openshift.io) 2025-12-04T11:51:54.262210841Z E1204 11:51:54.262171 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go@v1.5.2/tools/cache/reflector.go:243: Failed to watch *v1.Build: failed to list *v1.Build: the server is currently unable to handle the request (get builds.build.openshift.io)" 2025-12-04T11:52:01.787119147Z W1204 11:52:01.787010 1 reflector.go:561] k8s.io/client-go@v1.5.2/tools/cache/reflector.go:243: failed to list *v1.TemplateInstance: the server is currently unable to handle the request (get templateinstances.template.openshift.io) 2025-12-04T11:52:01.787119147Z E1204 11:52:01.787095 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go@v1.5.2/tools/cache/reflector.go:243: Failed to watch *v1.TemplateInstance: failed to list *v1.TemplateInstance: the server is currently unable to handle the request (get templateinstances.template.openshift.io)" 2025-12-04T11:52:05.706582533Z W1204 11:52:05.706499 1 reflector.go:561] k8s.io/client-go@v1.5.2/tools/cache/reflector.go:243: failed to list *v1.DeploymentConfig: the server is currently unable to handle the request (get deploymentconfigs.apps.openshift.io) 2025-12-04T11:52:05.706582533Z E1204 11:52:05.706559 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go@v1.5.2/tools/cache/reflector.go:243: Failed to watch *v1.DeploymentConfig: failed to list *v1.DeploymentConfig: the server is currently unable to handle the request (get deploymentconfigs.apps.openshift.io)" 2025-12-04T11:52:25.637249557Z W1204 11:52:25.637170 1 reflector.go:561] k8s.io/client-go@v1.5.2/tools/cache/reflector.go:243: failed to list *v1.ImageStream: the server is currently unable to handle the request (get imagestreams.image.openshift.io) 2025-12-04T11:52:25.637249557Z E1204 11:52:25.637228 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go@v1.5.2/tools/cache/reflector.go:243: Failed to watch *v1.ImageStream: failed to list *v1.ImageStream: the server is currently unable to handle the request (get imagestreams.image.openshift.io)" 2025-12-04T11:52:36.188094301Z W1204 11:52:36.188028 1 reflector.go:561] k8s.io/client-go@v1.5.2/tools/cache/reflector.go:243: failed to list *v1.Image: the server is currently unable to handle the request (get images.image.openshift.io) 2025-12-04T11:52:36.188220185Z E1204 11:52:36.188197 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go@v1.5.2/tools/cache/reflector.go:243: Failed to watch *v1.Image: failed to list *v1.Image: the server is currently unable to handle the request (get images.image.openshift.io)" 2025-12-04T11:52:36.274827560Z W1204 11:52:36.274742 1 reflector.go:561] k8s.io/client-go@v1.5.2/tools/cache/reflector.go:243: failed to list *v1.BuildConfig: the server is currently unable to handle the request (get buildconfigs.build.openshift.io) 2025-12-04T11:52:36.274883242Z E1204 11:52:36.274796 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go@v1.5.2/tools/cache/reflector.go:243: Failed to watch *v1.BuildConfig: failed to list *v1.BuildConfig: the server is currently unable to handle the request (get buildconfigs.build.openshift.io)" 2025-12-04T11:52:50.328425019Z W1204 11:52:50.328330 1 reflector.go:561] k8s.io/client-go@v1.5.2/tools/cache/reflector.go:243: failed to list *v1.DeploymentConfig: the server is currently unable to handle the request (get deploymentconfigs.apps.openshift.io) 2025-12-04T11:52:50.328425019Z E1204 11:52:50.328407 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go@v1.5.2/tools/cache/reflector.go:243: Failed to watch *v1.DeploymentConfig: failed to list *v1.DeploymentConfig: the server is currently unable to handle the request (get deploymentconfigs.apps.openshift.io)" 2025-12-04T11:52:50.611678685Z E1204 11:52:50.611600 1 leaderelection.go:436] error retrieving resource lock openshift-controller-manager/openshift-master-controllers: the server was unable to return a response in the time allotted, but may still be processing the request (get leases.coordination.k8s.io openshift-master-controllers) 2025-12-04T11:52:50.852269317Z W1204 11:52:50.852177 1 reflector.go:561] k8s.io/client-go@v1.5.2/tools/cache/reflector.go:243: failed to list *v1.Build: the server is currently unable to handle the request (get builds.build.openshift.io) 2025-12-04T11:52:50.852269317Z E1204 11:52:50.852248 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go@v1.5.2/tools/cache/reflector.go:243: Failed to watch *v1.Build: failed to list *v1.Build: the server is currently unable to handle the request (get builds.build.openshift.io)" 2025-12-04T11:52:53.778621711Z W1204 11:52:53.778532 1 reflector.go:561] k8s.io/client-go@v1.5.2/tools/cache/reflector.go:243: failed to list *v1.TemplateInstance: the server is currently unable to handle the request (get templateinstances.template.openshift.io) 2025-12-04T11:52:53.778621711Z E1204 11:52:53.778593 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go@v1.5.2/tools/cache/reflector.go:243: Failed to watch *v1.TemplateInstance: failed to list *v1.TemplateInstance: the server is currently unable to handle the request (get templateinstances.template.openshift.io)" 2025-12-04T11:52:56.034050630Z W1204 11:52:56.033949 1 reflector.go:561] k8s.io/client-go@v1.5.2/tools/cache/reflector.go:243: failed to list *v1.ImageStream: the server is currently unable to handle the request (get imagestreams.image.openshift.io) 2025-12-04T11:52:56.034050630Z E1204 11:52:56.034017 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go@v1.5.2/tools/cache/reflector.go:243: Failed to watch *v1.ImageStream: failed to list *v1.ImageStream: the server is currently unable to handle the request (get imagestreams.image.openshift.io)" 2025-12-04T11:53:03.607746863Z I1204 11:53:03.607609 1 leaderelection.go:297] failed to renew lease openshift-controller-manager/openshift-master-controllers: timed out waiting for the condition 2025-12-04T11:53:21.454310684Z W1204 11:53:21.454240 1 reflector.go:561] k8s.io/client-go@v1.5.2/tools/cache/reflector.go:243: failed to list *v1.BuildConfig: the server is currently unable to handle the request (get buildconfigs.build.openshift.io) 2025-12-04T11:53:21.454379706Z E1204 11:53:21.454307 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go@v1.5.2/tools/cache/reflector.go:243: Failed to watch *v1.BuildConfig: failed to list *v1.BuildConfig: the server is currently unable to handle the request (get buildconfigs.build.openshift.io)" 2025-12-04T11:53:22.395349782Z W1204 11:53:22.395278 1 reflector.go:561] k8s.io/client-go@v1.5.2/tools/cache/reflector.go:243: failed to list *v1.Build: the server is currently unable to handle the request (get builds.build.openshift.io) 2025-12-04T11:53:22.395512037Z E1204 11:53:22.395481 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go@v1.5.2/tools/cache/reflector.go:243: Failed to watch *v1.Build: failed to list *v1.Build: the server is currently unable to handle the request (get builds.build.openshift.io)" 2025-12-04T11:53:23.813267658Z W1204 11:53:23.813187 1 reflector.go:561] k8s.io/client-go@v1.5.2/tools/cache/reflector.go:243: failed to list *v1.Image: the server is currently unable to handle the request (get images.image.openshift.io) 2025-12-04T11:53:23.813325979Z E1204 11:53:23.813254 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go@v1.5.2/tools/cache/reflector.go:243: Failed to watch *v1.Image: failed to list *v1.Image: the server is currently unable to handle the request (get images.image.openshift.io)" 2025-12-04T11:53:28.391112556Z W1204 11:53:28.391043 1 reflector.go:561] k8s.io/client-go@v1.5.2/tools/cache/reflector.go:243: failed to list *v1.DeploymentConfig: the server is currently unable to handle the request (get deploymentconfigs.apps.openshift.io) 2025-12-04T11:53:28.391267121Z E1204 11:53:28.391246 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go@v1.5.2/tools/cache/reflector.go:243: Failed to watch *v1.DeploymentConfig: failed to list *v1.DeploymentConfig: the server is currently unable to handle the request (get deploymentconfigs.apps.openshift.io)" 2025-12-04T11:53:37.612709128Z E1204 11:53:37.612592 1 leaderelection.go:322] Failed to release lock: Timeout: request did not complete within requested timeout - context deadline exceeded 2025-12-04T11:53:37.612709128Z W1204 11:53:37.612696 1 controller_manager.go:108] Controller Manager received stop signal: leaderelection lost config/pod/openshift-controller-manager/controller-manager-6686654b8d-rrndk.json0000640000000000000000000001563115114274105026135 0ustar0000000000000000{"metadata":{"name":"controller-manager-6686654b8d-rrndk","generateName":"controller-manager-6686654b8d-","namespace":"openshift-controller-manager","uid":"24506aa4-ab78-49df-bb58-59093498f13d","resourceVersion":"11959","creationTimestamp":"2025-12-04T11:38:45Z","labels":{"app":"openshift-controller-manager-a","controller-manager":"true","openshift-route-controller-manager-anti-affinity":"true","pod-template-hash":"6686654b8d"},"annotations":{"configmaps/client-ca":"6701","configmaps/config":"9056","configmaps/openshift-global-ca":"4098","configmaps/openshift-service-ca":"5186","k8s.ovn.org/pod-networks":"{\"default\":{\"ip_addresses\":[\"10.128.0.69/23\"],\"mac_address\":\"0a:58:0a:80:00:45\",\"gateway_ips\":[\"10.128.0.1\"],\"routes\":[{\"dest\":\"10.128.0.0/16\",\"nextHop\":\"10.128.0.1\"},{\"dest\":\"172.30.0.0/16\",\"nextHop\":\"10.128.0.1\"},{\"dest\":\"169.254.0.5/32\",\"nextHop\":\"10.128.0.1\"},{\"dest\":\"100.64.0.0/16\",\"nextHop\":\"10.128.0.1\"}],\"ip_address\":\"10.128.0.69/23\",\"gateway_ip\":\"10.128.0.1\",\"role\":\"primary\"}}","k8s.v1.cni.cncf.io/network-status":"[{\n \"name\": \"ovn-kubernetes\",\n \"interface\": \"eth0\",\n \"ips\": [\n \"10.128.0.69\"\n ],\n \"mac\": \"0a:58:0a:80:00:45\",\n \"default\": true,\n \"dns\": {}\n}]","openshift.io/required-scc":"restricted-v2","openshift.io/scc":"restricted-v2","openshiftcontrollermanagers.operator.openshift.io/cluster":"2","seccomp.security.alpha.kubernetes.io/pod":"runtime/default"},"ownerReferences":[{"apiVersion":"apps/v1","kind":"ReplicaSet","name":"controller-manager-6686654b8d","uid":"b23ece4f-4ec8-4a7f-9262-2ff874940c8c","controller":true,"blockOwnerDeletion":true}]},"spec":{"volumes":[{"name":"config","configMap":{"name":"config","defaultMode":420}},{"name":"client-ca","configMap":{"name":"client-ca","defaultMode":420}},{"name":"serving-cert","secret":{"secretName":"serving-cert","defaultMode":420}},{"name":"proxy-ca-bundles","configMap":{"name":"openshift-global-ca","items":[{"key":"ca-bundle.crt","path":"tls-ca-bundle.pem"}],"defaultMode":420}},{"name":"kube-api-access-w2qnk","projected":{"sources":[{"serviceAccountToken":{"expirationSeconds":3607,"path":"token"}},{"configMap":{"name":"kube-root-ca.crt","items":[{"key":"ca.crt","path":"ca.crt"}]}},{"downwardAPI":{"items":[{"path":"namespace","fieldRef":{"apiVersion":"v1","fieldPath":"metadata.namespace"}}]}},{"configMap":{"name":"openshift-service-ca.crt","items":[{"key":"service-ca.crt","path":"service-ca.crt"}]}}],"defaultMode":420}}],"containers":[{"name":"controller-manager","image":"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:eddedae7578d79b5a3f748000ae5c00b9f14a04710f9f9ec7b52fc569be5dfb8","command":["openshift-controller-manager","start"],"args":["--config=/var/run/configmaps/config/config.yaml","-v=2"],"ports":[{"containerPort":8443,"protocol":"TCP"}],"resources":{"requests":{"cpu":"100m","memory":"100Mi"}},"volumeMounts":[{"name":"config","mountPath":"/var/run/configmaps/config"},{"name":"client-ca","mountPath":"/var/run/configmaps/client-ca"},{"name":"serving-cert","mountPath":"/var/run/secrets/serving-cert"},{"name":"proxy-ca-bundles","mountPath":"/etc/pki/ca-trust/extracted/pem"},{"name":"kube-api-access-w2qnk","readOnly":true,"mountPath":"/var/run/secrets/kubernetes.io/serviceaccount"}],"livenessProbe":{"httpGet":{"path":"healthz","port":8443,"scheme":"HTTPS"},"initialDelaySeconds":30,"timeoutSeconds":1,"periodSeconds":10,"successThreshold":1,"failureThreshold":3},"readinessProbe":{"httpGet":{"path":"healthz","port":8443,"scheme":"HTTPS"},"timeoutSeconds":1,"periodSeconds":10,"successThreshold":1,"failureThreshold":10},"terminationMessagePath":"/dev/termination-log","terminationMessagePolicy":"FallbackToLogsOnError","imagePullPolicy":"IfNotPresent","securityContext":{"capabilities":{"drop":["ALL"]},"runAsUser":1000570000,"allowPrivilegeEscalation":false}}],"restartPolicy":"Always","terminationGracePeriodSeconds":30,"dnsPolicy":"ClusterFirst","nodeSelector":{"node-role.kubernetes.io/master":""},"serviceAccountName":"openshift-controller-manager-sa","serviceAccount":"openshift-controller-manager-sa","nodeName":"master-0","securityContext":{"seLinuxOptions":{"level":"s0:c24,c9"},"runAsNonRoot":true,"fsGroup":1000570000,"seccompProfile":{"type":"RuntimeDefault"}},"affinity":{"podAntiAffinity":{"requiredDuringSchedulingIgnoredDuringExecution":[{"labelSelector":{"matchLabels":{"app":"openshift-controller-manager-a","controller-manager":"true","openshift-route-controller-manager-anti-affinity":"true"}},"topologyKey":"kubernetes.io/hostname"}]}},"schedulerName":"default-scheduler","tolerations":[{"key":"node-role.kubernetes.io/master","operator":"Exists","effect":"NoSchedule"},{"key":"node.kubernetes.io/unreachable","operator":"Exists","effect":"NoExecute","tolerationSeconds":120},{"key":"node.kubernetes.io/not-ready","operator":"Exists","effect":"NoExecute","tolerationSeconds":120},{"key":"node.kubernetes.io/memory-pressure","operator":"Exists","effect":"NoSchedule"}],"priorityClassName":"system-node-critical","priority":2000001000,"enableServiceLinks":true,"preemptionPolicy":"PreemptLowerPriority"},"status":{"phase":"Running","conditions":[{"type":"PodReadyToStartContainers","status":"True","lastProbeTime":null,"lastTransitionTime":"2025-12-04T11:45:12Z"},{"type":"Initialized","status":"True","lastProbeTime":null,"lastTransitionTime":"2025-12-04T11:38:48Z"},{"type":"Ready","status":"True","lastProbeTime":null,"lastTransitionTime":"2025-12-04T11:59:30Z"},{"type":"ContainersReady","status":"True","lastProbeTime":null,"lastTransitionTime":"2025-12-04T11:59:30Z"},{"type":"PodScheduled","status":"True","lastProbeTime":null,"lastTransitionTime":"2025-12-04T11:38:48Z"}],"hostIP":"192.168.32.10","hostIPs":[{"ip":"192.168.32.10"}],"podIP":"10.128.0.69","podIPs":[{"ip":"10.128.0.69"}],"startTime":"2025-12-04T11:38:48Z","containerStatuses":[{"name":"controller-manager","state":{"running":{"startedAt":"2025-12-04T11:53:38Z"}},"lastState":{"terminated":{"exitCode":0,"reason":"Completed","startedAt":"2025-12-04T11:45:11Z","finishedAt":"2025-12-04T11:53:37Z","containerID":"cri-o://89ac9be46243609bde604f4ef89b49eac3ee3ceca2042c8efeae7607ac835469"}},"ready":true,"restartCount":1,"image":"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:eddedae7578d79b5a3f748000ae5c00b9f14a04710f9f9ec7b52fc569be5dfb8","imageID":"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:eddedae7578d79b5a3f748000ae5c00b9f14a04710f9f9ec7b52fc569be5dfb8","containerID":"cri-o://c38ff2120777ff8b1be511702637cc316f617c906d6a31bbe9e628d01aaa9983","started":true,"volumeMounts":[{"name":"config","mountPath":"/var/run/configmaps/config"},{"name":"client-ca","mountPath":"/var/run/configmaps/client-ca"},{"name":"serving-cert","mountPath":"/var/run/secrets/serving-cert"},{"name":"proxy-ca-bundles","mountPath":"/etc/pki/ca-trust/extracted/pem"},{"name":"kube-api-access-w2qnk","mountPath":"/var/run/secrets/kubernetes.io/serviceaccount","readOnly":true,"recursiveReadOnly":"Disabled"}]}],"qosClass":"Burstable"}}openshift-controller-manager-operator_current.log0000640000000000000000000011733315114274105045254 0ustar0000000000000000config/pod/openshift-controller-manager-operator/logs/openshift-controller-manager-operator-6c8676f99d-7z9482025-12-04T11:59:29.931353926Z E1204 11:59:29.931256 1 base_controller.go:279] "Unhandled Error" err="OpenshiftControllerManagerStaticResources-StaticResources reconciliation failed: [\"assets/openshift-controller-manager/informer-clusterrole.yaml\" (string): Get \"https://172.30.0.1:443/apis/rbac.authorization.k8s.io/v1/clusterroles/system:openshift:openshift-controller-manager\": dial tcp 172.30.0.1:443: connect: connection refused, \"assets/openshift-controller-manager/informer-clusterrolebinding.yaml\" (string): Get \"https://172.30.0.1:443/apis/rbac.authorization.k8s.io/v1/clusterrolebindings/system:openshift:openshift-controller-manager\": dial tcp 172.30.0.1:443: connect: connection refused, \"assets/openshift-controller-manager/tokenreview-clusterrole.yaml\" (string): Get \"https://172.30.0.1:443/apis/rbac.authorization.k8s.io/v1/clusterroles/system:openshift:tokenreview-openshift-controller-manager\": dial tcp 172.30.0.1:443: connect: connection refused, \"assets/openshift-controller-manager/tokenreview-clusterrolebinding.yaml\" (string): Get \"https://172.30.0.1:443/apis/rbac.authorization.k8s.io/v1/clusterrolebindings/system:openshift:tokenreview-openshift-controller-manager\": dial tcp 172.30.0.1:443: connect: connection refused, \"assets/openshift-controller-manager/leader-role.yaml\" (string): Get \"https://172.30.0.1:443/apis/rbac.authorization.k8s.io/v1/namespaces/openshift-controller-manager/roles/system:openshift:leader-locking-openshift-controller-manager\": dial tcp 172.30.0.1:443: connect: connection refused, \"assets/openshift-controller-manager/leader-rolebinding.yaml\" (string): Get \"https://172.30.0.1:443/apis/rbac.authorization.k8s.io/v1/namespaces/openshift-controller-manager/rolebindings/system:openshift:leader-locking-openshift-controller-manager\": dial tcp 172.30.0.1:443: connect: connection refused, \"assets/openshift-controller-manager/ns.yaml\" (string): Get \"https://172.30.0.1:443/api/v1/namespaces/openshift-controller-manager\": dial tcp 172.30.0.1:443: connect: connection refused, \"assets/openshift-controller-manager/route-controller-manager-clusterrole.yaml\" (string): Get \"https://172.30.0.1:443/apis/rbac.authorization.k8s.io/v1/clusterroles/system:openshift:openshift-route-controller-manager\": dial tcp 172.30.0.1:443: connect: connection refused, \"assets/openshift-controller-manager/route-controller-manager-clusterrolebinding.yaml\" (string): Get \"https://172.30.0.1:443/apis/rbac.authorization.k8s.io/v1/clusterrolebindings/system:openshift:openshift-route-controller-manager\": dial tcp 172.30.0.1:443: connect: connection refused, \"assets/openshift-controller-manager/route-controller-manager-leader-role.yaml\" (string): Get \"https://172.30.0.1:443/apis/rbac.authorization.k8s.io/v1/namespaces/openshift-route-controller-manager/roles/system:openshift:leader-locking-openshift-route-controller-manager\": dial tcp 172.30.0.1:443: connect: connection refused, \"assets/openshift-controller-manager/route-controller-manager-leader-rolebinding.yaml\" (string): Get \"https://172.30.0.1:443/apis/rbac.authorization.k8s.io/v1/namespaces/openshift-route-controller-manager/rolebindings/system:openshift:leader-locking-openshift-route-controller-manager\": dial tcp 172.30.0.1:443: connect: connection refused, \"assets/openshift-controller-manager/route-controller-manager-ns.yaml\" (string): Get \"https://172.30.0.1:443/api/v1/namespaces/openshift-route-controller-manager\": dial tcp 172.30.0.1:443: connect: connection refused, \"assets/openshift-controller-manager/route-controller-manager-sa.yaml\" (string): Get \"https://172.30.0.1:443/api/v1/namespaces/openshift-route-controller-manager/serviceaccounts/route-controller-manager-sa\": dial tcp 172.30.0.1:443: connect: connection refused, \"assets/openshift-controller-manager/route-controller-manager-separate-sa-role.yaml\" (string): Get \"https://172.30.0.1:443/apis/rbac.authorization.k8s.io/v1/namespaces/openshift-infra/roles/system:openshift:sa-creating-route-controller-manager\": dial tcp 172.30.0.1:443: connect: connection refused, \"assets/openshift-controller-manager/route-controller-manager-separate-sa-rolebinding.yaml\" (string): Get \"https://172.30.0.1:443/apis/rbac.authorization.k8s.io/v1/namespaces/openshift-infra/rolebindings/system:openshift:sa-creating-route-controller-manager\": dial tcp 172.30.0.1:443: connect: connection refused, \"assets/openshift-controller-manager/route-controller-manager-servicemonitor-role.yaml\" (string): Get \"https://172.30.0.1:443/apis/rbac.authorization.k8s.io/v1/namespaces/openshift-route-controller-manager/roles/prometheus-k8s\": dial tcp 172.30.0.1:443: connect: connection refused, \"assets/openshift-controller-manager/route-controller-manager-servicemonitor-rolebinding.yaml\" (string): Get \"https://172.30.0.1:443/apis/rbac.authorization.k8s.io/v1/namespaces/openshift-route-controller-manager/rolebindings/prometheus-k8s\": dial tcp 172.30.0.1:443: connect: connection refused, \"assets/openshift-controller-manager/route-controller-manager-svc.yaml\" (string): Get \"https://172.30.0.1:443/api/v1/namespaces/openshift-route-controller-manager/services/route-controller-manager\": dial tcp 172.30.0.1:443: connect: connection refused, \"assets/openshift-controller-manager/route-controller-manager-tokenreview-clusterrole.yaml\" (string): Get \"https://172.30.0.1:443/apis/rbac.authorization.k8s.io/v1/clusterroles/system:openshift:tokenreview-openshift-route-controller-manager\": dial tcp 172.30.0.1:443: connect: connection refused, \"assets/openshift-controller-manager/route-controller-manager-tokenreview-clusterrolebinding.yaml\" (string): Get \"https://172.30.0.1:443/apis/rbac.authorization.k8s.io/v1/clusterrolebindings/system:openshift:tokenreview-openshift-route-controller-manager\": dial tcp 172.30.0.1:443: connect: connection refused, \"assets/openshift-controller-manager/route-controller-manager-ingress-to-route-controller-clusterrole.yaml\" (string): Get \"https://172.30.0.1:443/apis/rbac.authorization.k8s.io/v1/clusterroles/system:openshift:openshift-controller-manager:ingress-to-route-controller\": dial tcp 172.30.0.1:443: connect: connection refused, \"assets/openshift-controller-manager/route-controller-manager-ingress-to-route-controller-clusterrolebinding.yaml\" (string): Get \"https://172.30.0.1:443/apis/rbac.authorization.k8s.io/v1/clusterrolebindings/system:openshift:openshift-controller-manager:ingress-to-route-controller\": dial tcp 172.30.0.1:443: connect: connection refused, \"assets/openshift-controller-manager/old-leader-role.yaml\" (string): Get \"https://172.30.0.1:443/apis/rbac.authorization.k8s.io/v1/namespaces/kube-system/roles/system:openshift:leader-locking-openshift-controller-manager\": dial tcp 172.30.0.1:443: connect: connection refused, \"assets/openshift-controller-manager/old-leader-rolebinding.yaml\" (string): Get \"https://172.30.0.1:443/apis/rbac.authorization.k8s.io/v1/namespaces/kube-system/rolebindings/system:openshift:leader-locking-openshift-controller-manager\": dial tcp 172.30.0.1:443: connect: connection refused, \"assets/openshift-controller-manager/separate-sa-role.yaml\" (string): Get \"https://172.30.0.1:443/apis/rbac.authorization.k8s.io/v1/namespaces/openshift-infra/roles/system:openshift:sa-creating-openshift-controller-manager\": dial tcp 172.30.0.1:443: connect: connection refused, \"assets/openshift-controller-manager/separate-sa-rolebinding.yaml\" (string): Get \"https://172.30.0.1:443/apis/rbac.authorization.k8s.io/v1/namespaces/openshift-infra/rolebindings/system:openshift:sa-creating-openshift-controller-manager\": dial tcp 172.30.0.1:443: connect: connection refused, \"assets/openshift-controller-manager/sa.yaml\" (string): Get \"https://172.30.0.1:443/api/v1/namespaces/openshift-controller-manager/serviceaccounts/openshift-controller-manager-sa\": dial tcp 172.30.0.1:443: connect: connection refused, \"assets/openshift-controller-manager/svc.yaml\" (string): Get \"https://172.30.0.1:443/api/v1/namespaces/openshift-controller-manager/services/controller-manager\": dial tcp 172.30.0.1:443: connect: connection refused, \"assets/openshift-controller-manager/servicemonitor-role.yaml\" (string): Get \"https://172.30.0.1:443/apis/rbac.authorization.k8s.io/v1/namespaces/openshift-controller-manager/roles/prometheus-k8s\": dial tcp 172.30.0.1:443: connect: connection refused, \"assets/openshift-controller-manager/servicemonitor-rolebinding.yaml\" (string): Get \"https://172.30.0.1:443/apis/rbac.authorization.k8s.io/v1/namespaces/openshift-controller-manager/rolebindings/prometheus-k8s\": dial tcp 172.30.0.1:443: connect: connection refused, \"assets/openshift-controller-manager/buildconfigstatus-clusterrole.yaml\" (string): Get \"https://172.30.0.1:443/apis/rbac.authorization.k8s.io/v1/clusterroles/system:openshift:openshift-controller-manager:update-buildconfig-status\": dial tcp 172.30.0.1:443: connect: connection refused, \"assets/openshift-controller-manager/buildconfigstatus-clusterrolebinding.yaml\" (string): Get \"https://172.30.0.1:443/apis/rbac.authorization.k8s.io/v1/clusterrolebindings/system:openshift:openshift-controller-manager:update-buildconfig-status\": dial tcp 172.30.0.1:443: connect: connection refused, \"assets/openshift-controller-manager/deployer-clusterrole.yaml\" (string): Get \"https://172.30.0.1:443/apis/rbac.authorization.k8s.io/v1/clusterroles/system:deployer\": dial tcp 172.30.0.1:443: connect: connection refused, \"assets/openshift-controller-manager/deployer-clusterrolebinding.yaml\" (string): Get \"https://172.30.0.1:443/apis/rbac.authorization.k8s.io/v1/clusterrolebindings/system:deployer\": dial tcp 172.30.0.1:443: connect: connection refused, \"assets/openshift-controller-manager/image-trigger-controller-clusterrole.yaml\" (string): Get \"https://172.30.0.1:443/apis/rbac.authorization.k8s.io/v1/clusterroles/system:openshift:openshift-controller-manager:image-trigger-controller\": dial tcp 172.30.0.1:443: connect: connection refused, \"assets/openshift-controller-manager/image-trigger-controller-clusterrolebinding.yaml\" (string): Get \"https://172.30.0.1:443/apis/rbac.authorization.k8s.io/v1/clusterrolebindings/system:openshift:openshift-controller-manager:image-trigger-controller\": dial tcp 172.30.0.1:443: connect: connection refused, \"assets/openshift-controller-manager/leader-ingress-to-route-controller-role.yaml\" (string): Delete \"https://172.30.0.1:443/apis/rbac.authorization.k8s.io/v1/namespaces/openshift-route-controller-manager/roles/system:openshift:openshift-controller-manager:leader-locking-ingress-to-route-controller\": dial tcp 172.30.0.1:443: connect: connection refused, \"assets/openshift-controller-manager/leader-ingress-to-route-controller-rolebinding.yaml\" (string): Delete \"https://172.30.0.1:443/apis/rbac.authorization.k8s.io/v1/namespaces/openshift-route-controller-manager/rolebindings/system:openshift:openshift-controller-manager:leader-locking-ingress-to-route-controller\": dial tcp 172.30.0.1:443: connect: connection refused]" 2025-12-04T11:59:46.707946819Z I1204 11:59:46.707862 1 reflector.go:368] Caches populated for *v1.Secret from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243 2025-12-04T11:59:47.182274732Z I1204 11:59:47.182171 1 reflector.go:368] Caches populated for *v1.Pod from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243 2025-12-04T11:59:47.881077183Z I1204 11:59:47.880977 1 reflector.go:368] Caches populated for *v1.ServiceAccount from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243 2025-12-04T11:59:49.555120282Z I1204 11:59:49.555010 1 reflector.go:368] Caches populated for *v1.Deployment from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243 2025-12-04T11:59:49.822694297Z I1204 11:59:49.822606 1 reflector.go:368] Caches populated for *v1.ServiceAccount from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243 2025-12-04T11:59:50.068099309Z I1204 11:59:50.067993 1 reflector.go:368] Caches populated for *v1.ClusterOperator from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243 2025-12-04T11:59:50.912613426Z I1204 11:59:50.912521 1 reflector.go:368] Caches populated for *v1.Secret from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243 2025-12-04T11:59:51.132067847Z I1204 11:59:51.131529 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243 2025-12-04T11:59:51.423604420Z I1204 11:59:51.423504 1 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243 2025-12-04T11:59:51.666965688Z I1204 11:59:51.666882 1 reflector.go:368] Caches populated for *v1.Namespace from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243 2025-12-04T11:59:51.965904565Z I1204 11:59:51.965823 1 reflector.go:368] Caches populated for *v1.Network from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243 2025-12-04T11:59:52.047861611Z I1204 11:59:52.047726 1 reflector.go:368] Caches populated for *v1.Secret from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243 2025-12-04T11:59:52.411912271Z I1204 11:59:52.411831 1 reflector.go:368] Caches populated for *v1.Deployment from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243 2025-12-04T11:59:52.412064476Z I1204 11:59:52.412028 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243 2025-12-04T11:59:52.487426352Z I1204 11:59:52.487326 1 reflector.go:368] Caches populated for *v1.Role from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243 2025-12-04T11:59:52.795209650Z I1204 11:59:52.795127 1 reflector.go:368] Caches populated for *v1.ClusterVersion from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243 2025-12-04T11:59:53.011179610Z I1204 11:59:53.011075 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243 2025-12-04T11:59:53.226254931Z I1204 11:59:53.226148 1 reflector.go:368] Caches populated for *v1.Proxy from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243 2025-12-04T11:59:53.373531695Z I1204 11:59:53.373111 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243 2025-12-04T11:59:53.744602837Z I1204 11:59:53.744532 1 reflector.go:368] Caches populated for *v1.Secret from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243 2025-12-04T11:59:53.873110207Z I1204 11:59:53.872990 1 reflector.go:368] Caches populated for *v1.Secret from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243 2025-12-04T11:59:53.880768220Z I1204 11:59:53.880697 1 reflector.go:368] Caches populated for *v1.Secret from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243 2025-12-04T11:59:54.254556329Z I1204 11:59:54.254455 1 reflector.go:368] Caches populated for *v1.Secret from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243 2025-12-04T11:59:54.507686075Z I1204 11:59:54.507550 1 reflector.go:368] Caches populated for *v1.RoleBinding from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243 2025-12-04T11:59:54.553723343Z I1204 11:59:54.553608 1 reflector.go:368] Caches populated for *v1.Image from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243 2025-12-04T11:59:54.771838161Z I1204 11:59:54.771744 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243 2025-12-04T11:59:54.772311716Z I1204 11:59:54.772276 1 tlsconfig.go:181] "Loaded client CA" index=0 certName="client-ca::kube-system::extension-apiserver-authentication::client-ca-file,client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file" certDetail="\"admin-kubeconfig-signer\" [] issuer=\"\" (2025-12-04 11:20:50 +0000 UTC to 2035-12-02 11:20:50 +0000 UTC (now=2025-12-04 11:59:54.772250084 +0000 UTC))" 2025-12-04T11:59:54.772321306Z I1204 11:59:54.772309 1 tlsconfig.go:181] "Loaded client CA" index=1 certName="client-ca::kube-system::extension-apiserver-authentication::client-ca-file,client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file" certDetail="\"kubelet-signer\" [] issuer=\"\" (2025-12-04 11:26:53 +0000 UTC to 2025-12-05 11:26:53 +0000 UTC (now=2025-12-04 11:59:54.772294886 +0000 UTC))" 2025-12-04T11:59:54.772349377Z I1204 11:59:54.772327 1 tlsconfig.go:181] "Loaded client CA" index=2 certName="client-ca::kube-system::extension-apiserver-authentication::client-ca-file,client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file" certDetail="\"kube-control-plane-signer\" [] issuer=\"\" (2025-12-04 11:26:54 +0000 UTC to 2026-12-04 11:26:54 +0000 UTC (now=2025-12-04 11:59:54.772315626 +0000 UTC))" 2025-12-04T11:59:54.772349377Z I1204 11:59:54.772342 1 tlsconfig.go:181] "Loaded client CA" index=3 certName="client-ca::kube-system::extension-apiserver-authentication::client-ca-file,client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file" certDetail="\"kube-apiserver-to-kubelet-signer\" [] issuer=\"\" (2025-12-04 11:26:54 +0000 UTC to 2026-12-04 11:26:54 +0000 UTC (now=2025-12-04 11:59:54.772334797 +0000 UTC))" 2025-12-04T11:59:54.772373818Z I1204 11:59:54.772354 1 tlsconfig.go:181] "Loaded client CA" index=4 certName="client-ca::kube-system::extension-apiserver-authentication::client-ca-file,client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file" certDetail="\"kubelet-bootstrap-kubeconfig-signer\" [] issuer=\"\" (2025-12-04 11:26:52 +0000 UTC to 2035-12-02 11:26:52 +0000 UTC (now=2025-12-04 11:59:54.772347917 +0000 UTC))" 2025-12-04T11:59:54.772373818Z I1204 11:59:54.772368 1 tlsconfig.go:181] "Loaded client CA" index=5 certName="client-ca::kube-system::extension-apiserver-authentication::client-ca-file,client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file" certDetail="\"kube-csr-signer_@1764848267\" [] issuer=\"kubelet-signer\" (2025-12-04 11:37:46 +0000 UTC to 2025-12-05 11:26:53 +0000 UTC (now=2025-12-04 11:59:54.772360768 +0000 UTC))" 2025-12-04T11:59:54.772690308Z I1204 11:59:54.772382 1 tlsconfig.go:181] "Loaded client CA" index=6 certName="client-ca::kube-system::extension-apiserver-authentication::client-ca-file,client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file" certDetail="\"openshift-kube-apiserver-operator_node-system-admin-signer@1764848255\" [] issuer=\"\" (2025-12-04 11:37:34 +0000 UTC to 2026-12-04 11:37:35 +0000 UTC (now=2025-12-04 11:59:54.772373208 +0000 UTC))" 2025-12-04T11:59:54.772690308Z I1204 11:59:54.772394 1 tlsconfig.go:181] "Loaded client CA" index=7 certName="client-ca::kube-system::extension-apiserver-authentication::client-ca-file,client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file" certDetail="\"aggregator-signer\" [] issuer=\"\" (2025-12-04 11:26:52 +0000 UTC to 2025-12-05 11:26:52 +0000 UTC (now=2025-12-04 11:59:54.772386949 +0000 UTC))" 2025-12-04T11:59:54.772690308Z I1204 11:59:54.772551 1 tlsconfig.go:203] "Loaded serving cert" certName="serving-cert::/var/run/secrets/serving-cert/tls.crt::/var/run/secrets/serving-cert/tls.key" certDetail="\"metrics.openshift-controller-manager-operator.svc\" [serving] validServingFor=[metrics.openshift-controller-manager-operator.svc,metrics.openshift-controller-manager-operator.svc.cluster.local] issuer=\"openshift-service-serving-signer@1764848266\" (2025-12-04 11:37:52 +0000 UTC to 2027-12-04 11:37:53 +0000 UTC (now=2025-12-04 11:59:54.772535443 +0000 UTC))" 2025-12-04T11:59:54.772690308Z I1204 11:59:54.772644 1 named_certificates.go:53] "Loaded SNI cert" index=0 certName="self-signed loopback" certDetail="\"apiserver-loopback-client@1764849371\" [serving] validServingFor=[apiserver-loopback-client] issuer=\"apiserver-loopback-client-ca@1764849371\" (2025-12-04 10:56:11 +0000 UTC to 2026-12-04 10:56:11 +0000 UTC (now=2025-12-04 11:59:54.772632606 +0000 UTC))" 2025-12-04T11:59:55.257075240Z I1204 11:59:55.257005 1 reflector.go:368] Caches populated for *v1.Role from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243 2025-12-04T11:59:55.564630180Z I1204 11:59:55.564521 1 reflector.go:368] Caches populated for *v1.RoleBinding from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243 2025-12-04T11:59:55.703878120Z I1204 11:59:55.703769 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243 2025-12-04T11:59:55.964969269Z I1204 11:59:55.964869 1 reflector.go:368] Caches populated for *v1.Namespace from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243 2025-12-04T11:59:56.201243371Z I1204 11:59:56.201158 1 reflector.go:368] Caches populated for *v1.RoleBinding from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243 2025-12-04T11:59:56.338493008Z I1204 11:59:56.338386 1 reflector.go:368] Caches populated for *v1.FeatureGate from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243 2025-12-04T11:59:57.472606807Z I1204 11:59:57.472494 1 reflector.go:368] Caches populated for *v1.OpenShiftControllerManager from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243 2025-12-04T11:59:58.295368744Z I1204 11:59:58.295261 1 reflector.go:368] Caches populated for *v1.Role from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243 2025-12-04T11:59:58.732473308Z I1204 11:59:58.732378 1 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243 2025-12-04T11:59:58.900006034Z I1204 11:59:58.899900 1 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243 2025-12-04T11:59:59.050886302Z I1204 11:59:59.048524 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243 2025-12-04T11:59:59.119003930Z I1204 11:59:59.118543 1 reflector.go:368] Caches populated for *v1.RoleBinding from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243 2025-12-04T11:59:59.215009861Z I1204 11:59:59.214882 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243 2025-12-04T11:59:59.578933747Z I1204 11:59:59.576420 1 reflector.go:368] Caches populated for *v1.Build from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243 2025-12-04T12:00:00.519794875Z I1204 12:00:00.519679 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243 2025-12-04T12:00:01.321942460Z I1204 12:00:01.321841 1 reflector.go:368] Caches populated for *v1.Role from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243 2025-12-04T12:00:01.392449553Z I1204 12:00:01.392343 1 reflector.go:368] Caches populated for *v1.ClusterRole from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243 2025-12-04T12:00:01.773313385Z I1204 12:00:01.773228 1 reflector.go:368] Caches populated for *v1.ClusterRoleBinding from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243 2025-12-04T12:00:03.537416436Z I1204 12:00:03.537347 1 reflector.go:368] Caches populated for *v1.ServiceAccount from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243 2025-12-04T12:00:26.439184388Z I1204 12:00:26.438934 1 event.go:377] Event(v1.ObjectReference{Kind:"Deployment", Namespace:"openshift-controller-manager-operator", Name:"openshift-controller-manager-operator", UID:"0a7cfec2-7ffa-41a5-84c8-6e6d311b722b", APIVersion:"apps/v1", ResourceVersion:"", FieldPath:""}): type: 'Normal' reason: 'ObservedConfigChanged' Writing updated observed config:   map[string]any{ 2025-12-04T12:00:26.439184388Z    "build": map[string]any{"buildDefaults": map[string]any{"resources": map[string]any{}}, "imageTemplateFormat": map[string]any{"format": string("quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:31aa3c7464"...)}}, 2025-12-04T12:00:26.439184388Z    "controllers": []any{ 2025-12-04T12:00:26.439184388Z    ... // 8 identical elements 2025-12-04T12:00:26.439184388Z    string("openshift.io/deploymentconfig"), 2025-12-04T12:00:26.439184388Z    string("openshift.io/image-import"), 2025-12-04T12:00:26.439184388Z    strings.Join({ 2025-12-04T12:00:26.439184388Z +  "-", 2025-12-04T12:00:26.439184388Z    "openshift.io/image-puller-rolebindings", 2025-12-04T12:00:26.439184388Z    }, ""), 2025-12-04T12:00:26.439184388Z    string("openshift.io/image-signature-import"), 2025-12-04T12:00:26.439184388Z    string("openshift.io/image-trigger"), 2025-12-04T12:00:26.439184388Z    ... // 2 identical elements 2025-12-04T12:00:26.439184388Z    string("openshift.io/origin-namespace"), 2025-12-04T12:00:26.439184388Z    string("openshift.io/serviceaccount"), 2025-12-04T12:00:26.439184388Z    strings.Join({ 2025-12-04T12:00:26.439184388Z +  "-", 2025-12-04T12:00:26.439184388Z    "openshift.io/serviceaccount-pull-secrets", 2025-12-04T12:00:26.439184388Z    }, ""), 2025-12-04T12:00:26.439184388Z    string("openshift.io/templateinstance"), 2025-12-04T12:00:26.439184388Z    string("openshift.io/templateinstancefinalizer"), 2025-12-04T12:00:26.439184388Z    string("openshift.io/unidling"), 2025-12-04T12:00:26.439184388Z    }, 2025-12-04T12:00:26.439184388Z    "deployer": map[string]any{"imageTemplateFormat": map[string]any{"format": string("quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:42c3f5030d"...)}}, 2025-12-04T12:00:26.439184388Z    "featureGates": []any{string("BuildCSIVolumes=true")}, 2025-12-04T12:00:26.439184388Z    "ingress": map[string]any{"ingressIPNetworkCIDR": string("")}, 2025-12-04T12:00:26.439184388Z   } 2025-12-04T12:00:29.604261050Z I1204 12:00:29.603921 1 core.go:352] ConfigMap "openshift-controller-manager/config" changes: {"apiVersion":"v1","data":{"config.yaml":"{\"apiVersion\":\"openshiftcontrolplane.config.openshift.io/v1\",\"build\":{\"buildDefaults\":{\"resources\":{}},\"imageTemplateFormat\":{\"format\":\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:31aa3c746438e79d97c5978e8a01d5294af45fd3e331d58d9a50adaee75080a2\"}},\"controllers\":[\"openshift.io/build\",\"openshift.io/build-config-change\",\"openshift.io/builder-rolebindings\",\"openshift.io/builder-serviceaccount\",\"-openshift.io/default-rolebindings\",\"openshift.io/deployer\",\"openshift.io/deployer-rolebindings\",\"openshift.io/deployer-serviceaccount\",\"openshift.io/deploymentconfig\",\"openshift.io/image-import\",\"-openshift.io/image-puller-rolebindings\",\"openshift.io/image-signature-import\",\"openshift.io/image-trigger\",\"openshift.io/ingress-ip\",\"openshift.io/ingress-to-route\",\"openshift.io/origin-namespace\",\"openshift.io/serviceaccount\",\"-openshift.io/serviceaccount-pull-secrets\",\"openshift.io/templateinstance\",\"openshift.io/templateinstancefinalizer\",\"openshift.io/unidling\"],\"deployer\":{\"imageTemplateFormat\":{\"format\":\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:42c3f5030dd19d0c34c5bff920f0686a1cf6bf8f5d287109d566fc86d0cbde7f\"}},\"featureGates\":[\"BuildCSIVolumes=true\"],\"ingress\":{\"ingressIPNetworkCIDR\":\"\"},\"kind\":\"OpenShiftControllerManagerConfig\"}\n"},"kind":"ConfigMap","metadata":{"creationTimestamp":null,"managedFields":null,"resourceVersion":null,"uid":null}} 2025-12-04T12:00:29.604363483Z I1204 12:00:29.604321 1 event.go:377] Event(v1.ObjectReference{Kind:"Deployment", Namespace:"openshift-controller-manager-operator", Name:"openshift-controller-manager-operator", UID:"0a7cfec2-7ffa-41a5-84c8-6e6d311b722b", APIVersion:"apps/v1", ResourceVersion:"", FieldPath:""}): type: 'Normal' reason: 'ConfigMapUpdated' Updated ConfigMap/config -n openshift-controller-manager: 2025-12-04T12:00:29.604363483Z cause by changes in data.config.yaml 2025-12-04T12:00:29.622935851Z I1204 12:00:29.622571 1 core.go:352] ConfigMap "openshift-route-controller-manager/config" changes: {"apiVersion":"v1","data":{"config.yaml":"{\"apiVersion\":\"openshiftcontrolplane.config.openshift.io/v1\",\"build\":{\"buildDefaults\":{\"resources\":{}},\"imageTemplateFormat\":{\"format\":\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:31aa3c746438e79d97c5978e8a01d5294af45fd3e331d58d9a50adaee75080a2\"}},\"controllers\":[\"openshift.io/build\",\"openshift.io/build-config-change\",\"openshift.io/builder-rolebindings\",\"openshift.io/builder-serviceaccount\",\"-openshift.io/default-rolebindings\",\"openshift.io/deployer\",\"openshift.io/deployer-rolebindings\",\"openshift.io/deployer-serviceaccount\",\"openshift.io/deploymentconfig\",\"openshift.io/image-import\",\"-openshift.io/image-puller-rolebindings\",\"openshift.io/image-signature-import\",\"openshift.io/image-trigger\",\"openshift.io/ingress-ip\",\"openshift.io/ingress-to-route\",\"openshift.io/origin-namespace\",\"openshift.io/serviceaccount\",\"-openshift.io/serviceaccount-pull-secrets\",\"openshift.io/templateinstance\",\"openshift.io/templateinstancefinalizer\",\"openshift.io/unidling\"],\"deployer\":{\"imageTemplateFormat\":{\"format\":\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:42c3f5030dd19d0c34c5bff920f0686a1cf6bf8f5d287109d566fc86d0cbde7f\"}},\"featureGates\":[\"BuildCSIVolumes=true\"],\"ingress\":{\"ingressIPNetworkCIDR\":\"\"},\"kind\":\"OpenShiftControllerManagerConfig\",\"leaderElection\":{\"name\":\"openshift-route-controllers\"}}\n"},"kind":"ConfigMap","metadata":{"creationTimestamp":null,"managedFields":null,"resourceVersion":null,"uid":null}} 2025-12-04T12:00:29.622935851Z I1204 12:00:29.622750 1 event.go:377] Event(v1.ObjectReference{Kind:"Deployment", Namespace:"openshift-controller-manager-operator", Name:"openshift-controller-manager-operator", UID:"0a7cfec2-7ffa-41a5-84c8-6e6d311b722b", APIVersion:"apps/v1", ResourceVersion:"", FieldPath:""}): type: 'Normal' reason: 'ConfigMapUpdated' Updated ConfigMap/config -n openshift-route-controller-manager: 2025-12-04T12:00:29.622935851Z cause by changes in data.config.yaml 2025-12-04T12:00:29.635329154Z I1204 12:00:29.635225 1 apps.go:155] Deployment "openshift-controller-manager/controller-manager" changes: {"metadata":{"annotations":{"operator.openshift.io/spec-hash":"dcef3da2bf37bbb20a49ab3894d749dfc2f0230bd46f2e532404c65056f101b2"}},"spec":{"progressDeadlineSeconds":null,"revisionHistoryLimit":null,"template":{"metadata":{"annotations":{"configmaps/config":"13602","openshiftcontrollermanagers.operator.openshift.io/cluster":"3"}},"spec":{"containers":[{"args":["--config=/var/run/configmaps/config/config.yaml","-v=2"],"command":["openshift-controller-manager","start"],"image":"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:eddedae7578d79b5a3f748000ae5c00b9f14a04710f9f9ec7b52fc569be5dfb8","imagePullPolicy":"IfNotPresent","livenessProbe":{"httpGet":{"path":"healthz","port":8443,"scheme":"HTTPS"},"initialDelaySeconds":30},"name":"controller-manager","ports":[{"containerPort":8443}],"readinessProbe":{"failureThreshold":10,"httpGet":{"path":"healthz","port":8443,"scheme":"HTTPS"}},"resources":{"requests":{"cpu":"100m","memory":"100Mi"}},"securityContext":{"allowPrivilegeEscalation":false,"capabilities":{"drop":["ALL"]}},"terminationMessagePolicy":"FallbackToLogsOnError","volumeMounts":[{"mountPath":"/var/run/configmaps/config","name":"config"},{"mountPath":"/var/run/configmaps/client-ca","name":"client-ca"},{"mountPath":"/var/run/secrets/serving-cert","name":"serving-cert"},{"mountPath":"/etc/pki/ca-trust/extracted/pem","name":"proxy-ca-bundles"}]}],"dnsPolicy":null,"restartPolicy":null,"schedulerName":null,"serviceAccount":null,"terminationGracePeriodSeconds":null,"volumes":[{"configMap":{"name":"config"},"name":"config"},{"configMap":{"name":"client-ca"},"name":"client-ca"},{"name":"serving-cert","secret":{"secretName":"serving-cert"}},{"configMap":{"items":[{"key":"ca-bundle.crt","path":"tls-ca-bundle.pem"}],"name":"openshift-global-ca"},"name":"proxy-ca-bundles"}]}}}} 2025-12-04T12:00:29.647843420Z W1204 12:00:29.647730 1 warnings.go:70] spec.template.spec.nodeSelector[node-role.kubernetes.io/master]: use "node-role.kubernetes.io/control-plane" instead 2025-12-04T12:00:29.648415628Z I1204 12:00:29.648352 1 event.go:377] Event(v1.ObjectReference{Kind:"Deployment", Namespace:"openshift-controller-manager-operator", Name:"openshift-controller-manager-operator", UID:"0a7cfec2-7ffa-41a5-84c8-6e6d311b722b", APIVersion:"apps/v1", ResourceVersion:"", FieldPath:""}): type: 'Normal' reason: 'DeploymentUpdated' Updated Deployment.apps/controller-manager -n openshift-controller-manager because it changed 2025-12-04T12:00:29.657826667Z I1204 12:00:29.657694 1 apps.go:155] Deployment "openshift-route-controller-manager/route-controller-manager" changes: {"metadata":{"annotations":{"operator.openshift.io/spec-hash":"4c3fb7b63c5a659501751b6927f50c0134d57c4d6b010079bd4d1f51d81c5eda"}},"spec":{"progressDeadlineSeconds":null,"revisionHistoryLimit":null,"template":{"metadata":{"annotations":{"configmaps/config":"13604","openshiftcontrollermanagers.operator.openshift.io/cluster":"3"}},"spec":{"containers":[{"args":["--config=/var/run/configmaps/config/config.yaml","-v=2"],"command":["route-controller-manager","start"],"env":[{"name":"POD_NAME","valueFrom":{"fieldRef":{"fieldPath":"metadata.name"}}}],"image":"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c416b201d480bddb5a4960ec42f4740761a1335001cf84ba5ae19ad6857771b1","imagePullPolicy":"IfNotPresent","livenessProbe":{"httpGet":{"path":"healthz","port":8443,"scheme":"HTTPS"},"initialDelaySeconds":30},"name":"route-controller-manager","ports":[{"containerPort":8443}],"readinessProbe":{"failureThreshold":10,"httpGet":{"path":"healthz","port":8443,"scheme":"HTTPS"}},"resources":{"requests":{"cpu":"100m","memory":"100Mi"}},"securityContext":{"allowPrivilegeEscalation":false,"capabilities":{"drop":["ALL"]}},"terminationMessagePolicy":"FallbackToLogsOnError","volumeMounts":[{"mountPath":"/var/run/configmaps/config","name":"config"},{"mountPath":"/var/run/configmaps/client-ca","name":"client-ca"},{"mountPath":"/var/run/secrets/serving-cert","name":"serving-cert"}]}],"dnsPolicy":null,"restartPolicy":null,"schedulerName":null,"serviceAccount":null,"terminationGracePeriodSeconds":null,"volumes":[{"configMap":{"name":"config"},"name":"config"},{"configMap":{"name":"client-ca"},"name":"client-ca"},{"name":"serving-cert","secret":{"secretName":"serving-cert"}}]}}}} 2025-12-04T12:00:29.672172571Z W1204 12:00:29.672067 1 warnings.go:70] spec.template.spec.nodeSelector[node-role.kubernetes.io/master]: use "node-role.kubernetes.io/control-plane" instead 2025-12-04T12:00:29.672935045Z I1204 12:00:29.672854 1 event.go:377] Event(v1.ObjectReference{Kind:"Deployment", Namespace:"openshift-controller-manager-operator", Name:"openshift-controller-manager-operator", UID:"0a7cfec2-7ffa-41a5-84c8-6e6d311b722b", APIVersion:"apps/v1", ResourceVersion:"", FieldPath:""}): type: 'Normal' reason: 'DeploymentUpdated' Updated Deployment.apps/route-controller-manager -n openshift-route-controller-manager because it changed 2025-12-04T12:00:29.697117761Z I1204 12:00:29.697015 1 status_controller.go:225] clusteroperator/openshift-controller-manager diff {"status":{"conditions":[{"lastTransitionTime":"2025-12-04T11:37:45Z","message":"All is well","reason":"AsExpected","status":"False","type":"Degraded"},{"lastTransitionTime":"2025-12-04T11:37:49Z","message":"Progressing: deployment/controller-manager: observed generation is 5, desired generation is 6.\nProgressing: deployment/controller-manager: available replicas is 0, desired available replicas \u003e 1\nProgressing: deployment/route-controller-manager: observed generation is 5, desired generation is 6.\nProgressing: deployment/route-controller-manager: available replicas is 0, desired available replicas \u003e 1\nProgressing: openshiftcontrollermanagers.operator.openshift.io/cluster: observed generation is 2, desired generation is 3.","reason":"_DesiredStateNotYetAchieved","status":"True","type":"Progressing"},{"lastTransitionTime":"2025-12-04T11:50:40Z","message":"Available: no pods available on any node.","reason":"_NoPodsAvailable","status":"False","type":"Available"},{"lastTransitionTime":"2025-12-04T11:37:49Z","message":"All is well","reason":"AsExpected","status":"True","type":"Upgradeable"},{"lastTransitionTime":"2025-12-04T11:37:44Z","reason":"NoData","status":"Unknown","type":"EvaluationConditionsDetected"}]}} 2025-12-04T12:00:29.758744453Z I1204 12:00:29.757386 1 event.go:377] Event(v1.ObjectReference{Kind:"Deployment", Namespace:"openshift-controller-manager-operator", Name:"openshift-controller-manager-operator", UID:"0a7cfec2-7ffa-41a5-84c8-6e6d311b722b", APIVersion:"apps/v1", ResourceVersion:"", FieldPath:""}): type: 'Normal' reason: 'OperatorStatusChanged' Status for clusteroperator/openshift-controller-manager changed: Progressing message changed from "Progressing: deployment/controller-manager: available replicas is 0, desired available replicas > 1\nProgressing: deployment/route-controller-manager: available replicas is 0, desired available replicas > 1" to "Progressing: deployment/controller-manager: observed generation is 5, desired generation is 6.\nProgressing: deployment/controller-manager: available replicas is 0, desired available replicas > 1\nProgressing: deployment/route-controller-manager: observed generation is 5, desired generation is 6.\nProgressing: deployment/route-controller-manager: available replicas is 0, desired available replicas > 1\nProgressing: openshiftcontrollermanagers.operator.openshift.io/cluster: observed generation is 2, desired generation is 3." 2025-12-04T12:00:49.631221225Z I1204 12:00:49.631142 1 status_controller.go:225] clusteroperator/openshift-controller-manager diff {"status":{"conditions":[{"lastTransitionTime":"2025-12-04T11:37:45Z","message":"All is well","reason":"AsExpected","status":"False","type":"Degraded"},{"lastTransitionTime":"2025-12-04T11:37:49Z","message":"Progressing: deployment/controller-manager: observed generation is 5, desired generation is 6.\nProgressing: deployment/controller-manager: available replicas is 0, desired available replicas \u003e 1\nProgressing: deployment/route-controller-manager: observed generation is 5, desired generation is 6.\nProgressing: deployment/route-controller-manager: available replicas is 0, desired available replicas \u003e 1","reason":"_DesiredStateNotYetAchieved","status":"True","type":"Progressing"},{"lastTransitionTime":"2025-12-04T11:50:40Z","message":"Available: no pods available on any node.","reason":"_NoPodsAvailable","status":"False","type":"Available"},{"lastTransitionTime":"2025-12-04T11:37:49Z","message":"All is well","reason":"AsExpected","status":"True","type":"Upgradeable"},{"lastTransitionTime":"2025-12-04T11:37:44Z","reason":"NoData","status":"Unknown","type":"EvaluationConditionsDetected"}]}} 2025-12-04T12:00:49.644338601Z I1204 12:00:49.644212 1 event.go:377] Event(v1.ObjectReference{Kind:"Deployment", Namespace:"openshift-controller-manager-operator", Name:"openshift-controller-manager-operator", UID:"0a7cfec2-7ffa-41a5-84c8-6e6d311b722b", APIVersion:"apps/v1", ResourceVersion:"", FieldPath:""}): type: 'Normal' reason: 'OperatorStatusChanged' Status for clusteroperator/openshift-controller-manager changed: Progressing message changed from "Progressing: deployment/controller-manager: observed generation is 5, desired generation is 6.\nProgressing: deployment/controller-manager: available replicas is 0, desired available replicas > 1\nProgressing: deployment/route-controller-manager: observed generation is 5, desired generation is 6.\nProgressing: deployment/route-controller-manager: available replicas is 0, desired available replicas > 1\nProgressing: openshiftcontrollermanagers.operator.openshift.io/cluster: observed generation is 2, desired generation is 3." to "Progressing: deployment/controller-manager: observed generation is 5, desired generation is 6.\nProgressing: deployment/controller-manager: available replicas is 0, desired available replicas > 1\nProgressing: deployment/route-controller-manager: observed generation is 5, desired generation is 6.\nProgressing: deployment/route-controller-manager: available replicas is 0, desired available replicas > 1" openshift-controller-manager-operator_previous.log0000640000000000000000000004616715114274105045454 0ustar0000000000000000config/pod/openshift-controller-manager-operator/logs/openshift-controller-manager-operator-6c8676f99d-7z9482025-12-04T11:43:54.294843804Z I1204 11:43:54.294389 1 cmd.go:253] Using service-serving-cert provided certificates 2025-12-04T11:43:54.294843804Z I1204 11:43:54.294777 1 leaderelection.go:121] The leader election gives 4 retries and allows for 30s of clock skew. The kube-apiserver downtime tolerance is 78s. Worst non-graceful lease acquisition is 2m43s. Worst graceful lease acquisition is {26s}. 2025-12-04T11:43:54.298460373Z I1204 11:43:54.298411 1 observer_polling.go:159] Starting file observer 2025-12-04T11:44:54.304523392Z W1204 11:44:54.304412 1 builder.go:272] unable to get owner reference (falling back to namespace): the server was unable to return a response in the time allotted, but may still be processing the request (get pods openshift-controller-manager-operator-6c8676f99d-7z948) 2025-12-04T11:44:54.304889176Z I1204 11:44:54.304843 1 builder.go:304] openshift-controller-manager-operator version 4.18.0-202511181540.p2.gf1711cf.assembly.stream.el9-f1711cf-f1711cf30f683ec0eaa187cd5168caae9e8c1254 2025-12-04T11:44:54.629107866Z I1204 11:44:54.629001 1 secure_serving.go:57] Forcing use of http/1.1 only 2025-12-04T11:44:54.629107866Z W1204 11:44:54.629067 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected. 2025-12-04T11:44:54.629107866Z W1204 11:44:54.629072 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected. 2025-12-04T11:44:54.629107866Z W1204 11:44:54.629078 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected. 2025-12-04T11:44:54.629107866Z W1204 11:44:54.629083 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected. 2025-12-04T11:44:54.629107866Z W1204 11:44:54.629086 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected. 2025-12-04T11:44:54.629107866Z W1204 11:44:54.629090 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected. 2025-12-04T11:44:54.632334959Z I1204 11:44:54.632298 1 secure_serving.go:213] Serving securely on [::]:8443 2025-12-04T11:44:54.632465664Z I1204 11:44:54.632452 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController 2025-12-04T11:44:54.632507306Z I1204 11:44:54.632497 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController 2025-12-04T11:44:54.632581148Z I1204 11:44:54.632563 1 dynamic_serving_content.go:135] "Starting controller" name="serving-cert::/var/run/secrets/serving-cert/tls.crt::/var/run/secrets/serving-cert/tls.key" 2025-12-04T11:44:54.632732603Z I1204 11:44:54.632718 1 tlsconfig.go:243] "Starting DynamicServingCertificateController" 2025-12-04T11:44:54.632905189Z I1204 11:44:54.632886 1 configmap_cafile_content.go:205] "Starting controller" name="client-ca::kube-system::extension-apiserver-authentication::client-ca-file" 2025-12-04T11:44:54.632943420Z I1204 11:44:54.632932 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file 2025-12-04T11:44:54.633182459Z I1204 11:44:54.633122 1 configmap_cafile_content.go:205] "Starting controller" name="client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file" 2025-12-04T11:44:54.633221270Z I1204 11:44:54.633200 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file 2025-12-04T11:44:54.733405307Z I1204 11:44:54.733312 1 shared_informer.go:320] Caches are synced for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file 2025-12-04T11:44:54.733478818Z I1204 11:44:54.733390 1 shared_informer.go:320] Caches are synced for client-ca::kube-system::extension-apiserver-authentication::client-ca-file 2025-12-04T11:44:54.733478818Z I1204 11:44:54.733413 1 shared_informer.go:320] Caches are synced for RequestHeaderAuthRequestController 2025-12-04T11:45:14.630119381Z I1204 11:45:14.629659 1 builder.go:446] detected SingleReplicaTopologyMode, the original leader election has been altered for the default SingleReplicaTopology 2025-12-04T11:45:14.631841987Z I1204 11:45:14.630176 1 leaderelection.go:254] attempting to acquire leader lease openshift-controller-manager-operator/openshift-controller-manager-operator-lock... 2025-12-04T11:50:40.674166408Z I1204 11:50:40.674071 1 leaderelection.go:268] successfully acquired lease openshift-controller-manager-operator/openshift-controller-manager-operator-lock 2025-12-04T11:50:40.674267431Z I1204 11:50:40.674198 1 event.go:377] Event(v1.ObjectReference{Kind:"Lease", Namespace:"openshift-controller-manager-operator", Name:"openshift-controller-manager-operator-lock", UID:"54d43753-3a2b-4a23-9800-f9f22731e7cd", APIVersion:"coordination.k8s.io/v1", ResourceVersion:"11004", FieldPath:""}): type: 'Normal' reason: 'LeaderElection' openshift-controller-manager-operator-6c8676f99d-7z948_8b885916-2eb8-4873-93e8-1f63051ec480 became leader 2025-12-04T11:50:40.676677433Z I1204 11:50:40.675238 1 simple_featuregate_reader.go:171] Starting feature-gate-detector 2025-12-04T11:50:40.678450568Z I1204 11:50:40.678396 1 starter.go:117] FeatureGates initialized: knownFeatureGates=[AWSEFSDriverVolumeMetrics AdminNetworkPolicy AlibabaPlatform AzureWorkloadIdentity BareMetalLoadBalancer BuildCSIVolumes ChunkSizeMiB CloudDualStackNodeIPs DisableKubeletCloudCredentialProviders GCPLabelsTags HardwareSpeed IngressControllerLBSubnetsAWS KMSv1 ManagedBootImages ManagedBootImagesAWS MultiArchInstallAWS MultiArchInstallGCP NetworkDiagnosticsConfig NetworkLiveMigration NetworkSegmentation NewOLM NodeDisruptionPolicy OnClusterBuild PersistentIPsForVirtualization PrivateHostedZoneAWS SetEIPForNLBIngressController VSphereControlPlaneMachineSet VSphereDriverConfiguration VSphereMultiVCenters VSphereStaticIPs ValidatingAdmissionPolicy AWSClusterHostedDNS AdditionalRoutingCapabilities AutomatedEtcdBackup BootcNodeManagement CSIDriverSharedResource ClusterAPIInstall ClusterAPIInstallIBMCloud ClusterMonitoringConfig ConsolePluginContentSecurityPolicy DNSNameResolver DynamicResourceAllocation EtcdBackendQuota EventedPLEG Example ExternalOIDC ExternalOIDCWithUIDAndExtraClaimMappings GCPClusterHostedDNS GatewayAPI ImageStreamImportMode IngressControllerDynamicConfigurationManager InsightsConfig InsightsConfigAPI InsightsOnDemandDataGather InsightsRuntimeExtractor MachineAPIMigration MachineAPIOperatorDisableMachineHealthCheckController MachineAPIProviderOpenStack MachineConfigNodes MaxUnavailableStatefulSet MetricsCollectionProfiles MinimumKubeletVersion MixedCPUsAllocation MultiArchInstallAzure NodeSwap NutanixMultiSubnets OVNObservability OpenShiftPodSecurityAdmission PinnedImages PlatformOperators ProcMountType RouteAdvertisements RouteExternalCertificate ServiceAccountTokenNodeBinding SignatureStores SigstoreImageVerification StreamingCollectionEncodingToJSON StreamingCollectionEncodingToProtobuf TranslateStreamCloseWebsocketRequests UpgradeStatus UserNamespacesPodSecurityStandards UserNamespacesSupport VSphereMultiNetworks VolumeAttributesClass VolumeGroupSnapshot] 2025-12-04T11:50:40.678547061Z I1204 11:50:40.678505 1 event.go:377] Event(v1.ObjectReference{Kind:"Namespace", Namespace:"openshift-controller-manager-operator", Name:"openshift-controller-manager-operator", UID:"", APIVersion:"v1", ResourceVersion:"", FieldPath:""}): type: 'Normal' reason: 'FeatureGatesInitialized' FeatureGates updated to featuregates.Features{Enabled:[]v1.FeatureGateName{"AWSEFSDriverVolumeMetrics", "AdminNetworkPolicy", "AlibabaPlatform", "AzureWorkloadIdentity", "BareMetalLoadBalancer", "BuildCSIVolumes", "ChunkSizeMiB", "CloudDualStackNodeIPs", "DisableKubeletCloudCredentialProviders", "GCPLabelsTags", "HardwareSpeed", "IngressControllerLBSubnetsAWS", "KMSv1", "ManagedBootImages", "ManagedBootImagesAWS", "MultiArchInstallAWS", "MultiArchInstallGCP", "NetworkDiagnosticsConfig", "NetworkLiveMigration", "NetworkSegmentation", "NewOLM", "NodeDisruptionPolicy", "OnClusterBuild", "PersistentIPsForVirtualization", "PrivateHostedZoneAWS", "SetEIPForNLBIngressController", "VSphereControlPlaneMachineSet", "VSphereDriverConfiguration", "VSphereMultiVCenters", "VSphereStaticIPs", "ValidatingAdmissionPolicy"}, Disabled:[]v1.FeatureGateName{"AWSClusterHostedDNS", "AdditionalRoutingCapabilities", "AutomatedEtcdBackup", "BootcNodeManagement", "CSIDriverSharedResource", "ClusterAPIInstall", "ClusterAPIInstallIBMCloud", "ClusterMonitoringConfig", "ConsolePluginContentSecurityPolicy", "DNSNameResolver", "DynamicResourceAllocation", "EtcdBackendQuota", "EventedPLEG", "Example", "ExternalOIDC", "ExternalOIDCWithUIDAndExtraClaimMappings", "GCPClusterHostedDNS", "GatewayAPI", "ImageStreamImportMode", "IngressControllerDynamicConfigurationManager", "InsightsConfig", "InsightsConfigAPI", "InsightsOnDemandDataGather", "InsightsRuntimeExtractor", "MachineAPIMigration", "MachineAPIOperatorDisableMachineHealthCheckController", "MachineAPIProviderOpenStack", "MachineConfigNodes", "MaxUnavailableStatefulSet", "MetricsCollectionProfiles", "MinimumKubeletVersion", "MixedCPUsAllocation", "MultiArchInstallAzure", "NodeSwap", "NutanixMultiSubnets", "OVNObservability", "OpenShiftPodSecurityAdmission", "PinnedImages", "PlatformOperators", "ProcMountType", "RouteAdvertisements", "RouteExternalCertificate", "ServiceAccountTokenNodeBinding", "SignatureStores", "SigstoreImageVerification", "StreamingCollectionEncodingToJSON", "StreamingCollectionEncodingToProtobuf", "TranslateStreamCloseWebsocketRequests", "UpgradeStatus", "UserNamespacesPodSecurityStandards", "UserNamespacesSupport", "VSphereMultiNetworks", "VolumeAttributesClass", "VolumeGroupSnapshot"}} 2025-12-04T11:50:40.689324016Z I1204 11:50:40.689251 1 base_controller.go:76] Waiting for caches to sync for ImagePullSecretCleanupController 2025-12-04T11:50:40.689969765Z I1204 11:50:40.689915 1 operator.go:145] Starting OpenShiftControllerManagerOperator 2025-12-04T11:50:40.692207663Z I1204 11:50:40.690529 1 base_controller.go:76] Waiting for caches to sync for LoggingSyncer 2025-12-04T11:50:40.692207663Z I1204 11:50:40.690948 1 base_controller.go:76] Waiting for caches to sync for ConfigObserver 2025-12-04T11:50:40.692499792Z I1204 11:50:40.690977 1 base_controller.go:76] Waiting for caches to sync for UserCAObservationController 2025-12-04T11:50:40.692499792Z I1204 11:50:40.690989 1 base_controller.go:76] Waiting for caches to sync for openshift-controller-manager 2025-12-04T11:50:40.694284615Z I1204 11:50:40.694217 1 base_controller.go:76] Waiting for caches to sync for StatusSyncer_openshift-controller-manager 2025-12-04T11:50:40.700763331Z I1204 11:50:40.700482 1 base_controller.go:76] Waiting for caches to sync for OpenshiftControllerManagerStaticResources-StaticResources 2025-12-04T11:50:40.789386995Z I1204 11:50:40.789328 1 base_controller.go:82] Caches are synced for ImagePullSecretCleanupController 2025-12-04T11:50:40.789457337Z I1204 11:50:40.789446 1 base_controller.go:119] Starting #1 worker of ImagePullSecretCleanupController controller ... 2025-12-04T11:50:40.792237601Z I1204 11:50:40.792163 1 base_controller.go:82] Caches are synced for LoggingSyncer 2025-12-04T11:50:40.792237601Z I1204 11:50:40.792220 1 base_controller.go:119] Starting #1 worker of LoggingSyncer controller ... 2025-12-04T11:50:40.792281042Z I1204 11:50:40.792264 1 base_controller.go:82] Caches are synced for ConfigObserver 2025-12-04T11:50:40.792281042Z I1204 11:50:40.792277 1 base_controller.go:119] Starting #1 worker of ConfigObserver controller ... 2025-12-04T11:50:40.792476668Z I1204 11:50:40.792462 1 base_controller.go:82] Caches are synced for openshift-controller-manager 2025-12-04T11:50:40.792511119Z I1204 11:50:40.792502 1 base_controller.go:119] Starting #1 worker of openshift-controller-manager controller ... 2025-12-04T11:50:40.792574191Z I1204 11:50:40.792565 1 base_controller.go:82] Caches are synced for UserCAObservationController 2025-12-04T11:50:40.792597422Z I1204 11:50:40.792589 1 base_controller.go:119] Starting #1 worker of UserCAObservationController controller ... 2025-12-04T11:50:40.794914761Z I1204 11:50:40.794852 1 base_controller.go:82] Caches are synced for StatusSyncer_openshift-controller-manager 2025-12-04T11:50:40.794962543Z I1204 11:50:40.794949 1 base_controller.go:119] Starting #1 worker of StatusSyncer_openshift-controller-manager controller ... 2025-12-04T11:50:40.801536012Z I1204 11:50:40.801474 1 base_controller.go:82] Caches are synced for OpenshiftControllerManagerStaticResources-StaticResources 2025-12-04T11:50:40.801536012Z I1204 11:50:40.801508 1 base_controller.go:119] Starting #1 worker of OpenshiftControllerManagerStaticResources-StaticResources controller ... 2025-12-04T11:50:40.808143641Z I1204 11:50:40.808085 1 status_controller.go:225] clusteroperator/openshift-controller-manager diff {"status":{"conditions":[{"lastTransitionTime":"2025-12-04T11:37:45Z","message":"All is well","reason":"AsExpected","status":"False","type":"Degraded"},{"lastTransitionTime":"2025-12-04T11:37:49Z","message":"Progressing: deployment/controller-manager: available replicas is 0, desired available replicas \u003e 1\nProgressing: deployment/route-controller-manager: available replicas is 0, desired available replicas \u003e 1","reason":"_DesiredStateNotYetAchieved","status":"True","type":"Progressing"},{"lastTransitionTime":"2025-12-04T11:50:40Z","message":"Available: no pods available on any node.","reason":"_NoPodsAvailable","status":"False","type":"Available"},{"lastTransitionTime":"2025-12-04T11:37:49Z","message":"All is well","reason":"AsExpected","status":"True","type":"Upgradeable"},{"lastTransitionTime":"2025-12-04T11:37:44Z","reason":"NoData","status":"Unknown","type":"EvaluationConditionsDetected"}]}} 2025-12-04T11:50:40.817398870Z I1204 11:50:40.816839 1 event.go:377] Event(v1.ObjectReference{Kind:"Namespace", Namespace:"openshift-controller-manager-operator", Name:"openshift-controller-manager-operator", UID:"", APIVersion:"v1", ResourceVersion:"", FieldPath:""}): type: 'Normal' reason: 'OperatorStatusChanged' Status for clusteroperator/openshift-controller-manager changed: Progressing message changed from "Progressing: deployment/controller-manager: observed generation is 4, desired generation is 5.\nProgressing: deployment/route-controller-manager: observed generation is 4, desired generation is 5." to "Progressing: deployment/controller-manager: available replicas is 0, desired available replicas > 1\nProgressing: deployment/route-controller-manager: available replicas is 0, desired available replicas > 1",Available changed from True to False ("Available: no pods available on any node.") 2025-12-04T11:52:14.683870085Z E1204 11:52:14.683718 1 leaderelection.go:429] Failed to update lock optimitically: Timeout: request did not complete within requested timeout - context deadline exceeded, falling back to slow path 2025-12-04T11:52:40.795630548Z E1204 11:52:40.795550 1 base_controller.go:279] "Unhandled Error" err="openshift-controller-manager reconciliation failed: unable to get operator configuration: the server was unable to return a response in the time allotted, but may still be processing the request (get openshiftcontrollermanagers.operator.openshift.io cluster)" 2025-12-04T11:52:40.795630548Z E1204 11:52:40.795601 1 base_controller.go:279] "Unhandled Error" err="ConfigObserver reconciliation failed: unable to get operator configuration: the server was unable to return a response in the time allotted, but may still be processing the request (get openshiftcontrollermanagers.operator.openshift.io cluster)" 2025-12-04T11:53:14.686207189Z E1204 11:53:14.686092 1 leaderelection.go:436] error retrieving resource lock openshift-controller-manager-operator/openshift-controller-manager-operator-lock: the server was unable to return a response in the time allotted, but may still be processing the request (get leases.coordination.k8s.io openshift-controller-manager-operator-lock) 2025-12-04T11:53:40.819669153Z E1204 11:53:40.819570 1 base_controller.go:279] "Unhandled Error" err="openshift-controller-manager reconciliation failed: unable to get operator configuration: the server was unable to return a response in the time allotted, but may still be processing the request (get openshiftcontrollermanagers.operator.openshift.io cluster)" 2025-12-04T11:53:40.819747875Z E1204 11:53:40.819706 1 base_controller.go:279] "Unhandled Error" err="ConfigObserver reconciliation failed: unable to get operator configuration: the server was unable to return a response in the time allotted, but may still be processing the request (get openshiftcontrollermanagers.operator.openshift.io cluster)" 2025-12-04T11:54:40.823114010Z E1204 11:54:40.822927 1 base_controller.go:279] "Unhandled Error" err="ConfigObserver reconciliation failed: unable to get operator configuration: the server was unable to return a response in the time allotted, but may still be processing the request (get openshiftcontrollermanagers.operator.openshift.io cluster)" 2025-12-04T11:54:40.823114010Z E1204 11:54:40.823007 1 base_controller.go:279] "Unhandled Error" err="openshift-controller-manager reconciliation failed: unable to get operator configuration: the server was unable to return a response in the time allotted, but may still be processing the request (get openshiftcontrollermanagers.operator.openshift.io cluster)" 2025-12-04T11:54:48.690716177Z E1204 11:54:48.690614 1 leaderelection.go:429] Failed to update lock optimitically: Timeout: request did not complete within requested timeout - context deadline exceeded, falling back to slow path 2025-12-04T11:55:40.680540088Z E1204 11:55:40.680407 1 leaderelection.go:436] error retrieving resource lock openshift-controller-manager-operator/openshift-controller-manager-operator-lock: Get "https://172.30.0.1:443/apis/coordination.k8s.io/v1/namespaces/openshift-controller-manager-operator/leases/openshift-controller-manager-operator-lock?timeout=4m0s": context deadline exceeded 2025-12-04T11:55:40.680540088Z I1204 11:55:40.680487 1 leaderelection.go:297] failed to renew lease openshift-controller-manager-operator/openshift-controller-manager-operator-lock: timed out waiting for the condition 2025-12-04T11:55:40.825694906Z E1204 11:55:40.825589 1 base_controller.go:279] "Unhandled Error" err="openshift-controller-manager reconciliation failed: unable to get operator configuration: the server was unable to return a response in the time allotted, but may still be processing the request (get openshiftcontrollermanagers.operator.openshift.io cluster)" 2025-12-04T11:55:40.825694906Z E1204 11:55:40.825662 1 base_controller.go:279] "Unhandled Error" err="ConfigObserver reconciliation failed: unable to get operator configuration: the server was unable to return a response in the time allotted, but may still be processing the request (get openshiftcontrollermanagers.operator.openshift.io cluster)" 2025-12-04T11:55:46.781371859Z W1204 11:55:46.781309 1 leaderelection.go:84] leader election lost openshift-controller-manager-operator-6c8676f99d-7z948.json0000640000000000000000000001435315114274105033313 0ustar0000000000000000config/pod/openshift-controller-manager-operator{"metadata":{"name":"openshift-controller-manager-operator-6c8676f99d-7z948","generateName":"openshift-controller-manager-operator-6c8676f99d-","namespace":"openshift-controller-manager-operator","uid":"3322cc5a-f1f7-4522-b423-19fb7f38cd43","resourceVersion":"11269","creationTimestamp":"2025-12-04T11:35:49Z","labels":{"app":"openshift-controller-manager-operator","pod-template-hash":"6c8676f99d"},"annotations":{"k8s.ovn.org/pod-networks":"{\"default\":{\"ip_addresses\":[\"10.128.0.8/23\"],\"mac_address\":\"0a:58:0a:80:00:08\",\"gateway_ips\":[\"10.128.0.1\"],\"routes\":[{\"dest\":\"10.128.0.0/16\",\"nextHop\":\"10.128.0.1\"},{\"dest\":\"172.30.0.0/16\",\"nextHop\":\"10.128.0.1\"},{\"dest\":\"169.254.0.5/32\",\"nextHop\":\"10.128.0.1\"},{\"dest\":\"100.64.0.0/16\",\"nextHop\":\"10.128.0.1\"}],\"ip_address\":\"10.128.0.8/23\",\"gateway_ip\":\"10.128.0.1\",\"role\":\"primary\"}}","k8s.v1.cni.cncf.io/network-status":"[{\n \"name\": \"ovn-kubernetes\",\n \"interface\": \"eth0\",\n \"ips\": [\n \"10.128.0.8\"\n ],\n \"mac\": \"0a:58:0a:80:00:08\",\n \"default\": true,\n \"dns\": {}\n}]","openshift.io/required-scc":"nonroot-v2","openshift.io/scc":"nonroot-v2","seccomp.security.alpha.kubernetes.io/pod":"runtime/default"},"ownerReferences":[{"apiVersion":"apps/v1","kind":"ReplicaSet","name":"openshift-controller-manager-operator-6c8676f99d","uid":"11692e26-0986-4122-ac2a-530fcf6c0e46","controller":true,"blockOwnerDeletion":true}]},"spec":{"volumes":[{"name":"serving-cert","secret":{"secretName":"openshift-controller-manager-operator-serving-cert","defaultMode":420,"optional":true}},{"name":"config","configMap":{"name":"openshift-controller-manager-operator-config","defaultMode":420}},{"name":"kube-api-access-k449s","projected":{"sources":[{"serviceAccountToken":{"expirationSeconds":3607,"path":"token"}},{"configMap":{"name":"kube-root-ca.crt","items":[{"key":"ca.crt","path":"ca.crt"}]}},{"downwardAPI":{"items":[{"path":"namespace","fieldRef":{"apiVersion":"v1","fieldPath":"metadata.namespace"}}]}},{"configMap":{"name":"openshift-service-ca.crt","items":[{"key":"service-ca.crt","path":"service-ca.crt"}]}}],"defaultMode":420}}],"containers":[{"name":"openshift-controller-manager-operator","image":"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8eabac819f289e29d75c7ab172d8124554849a47f0b00770928c3eb19a5a31c4","command":["cluster-openshift-controller-manager-operator","operator"],"args":["--config=/var/run/configmaps/config/config.yaml"],"ports":[{"name":"metrics","containerPort":8443,"protocol":"TCP"}],"env":[{"name":"RELEASE_VERSION","value":"4.18.29"},{"name":"IMAGE","value":"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:eddedae7578d79b5a3f748000ae5c00b9f14a04710f9f9ec7b52fc569be5dfb8"},{"name":"OPERATOR_IMAGE_VERSION","value":"4.18.29"},{"name":"ROUTE_CONTROLLER_MANAGER_IMAGE","value":"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c416b201d480bddb5a4960ec42f4740761a1335001cf84ba5ae19ad6857771b1"},{"name":"POD_NAME","valueFrom":{"fieldRef":{"apiVersion":"v1","fieldPath":"metadata.name"}}}],"resources":{"requests":{"cpu":"10m","memory":"50Mi"}},"volumeMounts":[{"name":"config","mountPath":"/var/run/configmaps/config"},{"name":"serving-cert","mountPath":"/var/run/secrets/serving-cert"},{"name":"kube-api-access-k449s","readOnly":true,"mountPath":"/var/run/secrets/kubernetes.io/serviceaccount"}],"terminationMessagePath":"/dev/termination-log","terminationMessagePolicy":"FallbackToLogsOnError","imagePullPolicy":"IfNotPresent","securityContext":{"capabilities":{"drop":["ALL"]},"allowPrivilegeEscalation":false}}],"restartPolicy":"Always","terminationGracePeriodSeconds":30,"dnsPolicy":"ClusterFirst","nodeSelector":{"node-role.kubernetes.io/master":""},"serviceAccountName":"openshift-controller-manager-operator","serviceAccount":"openshift-controller-manager-operator","nodeName":"master-0","securityContext":{"seLinuxOptions":{"level":"s0:c15,c10"},"runAsUser":65534,"runAsNonRoot":true,"seccompProfile":{"type":"RuntimeDefault"}},"schedulerName":"default-scheduler","tolerations":[{"key":"node-role.kubernetes.io/master","operator":"Exists","effect":"NoSchedule"},{"key":"node.kubernetes.io/unreachable","operator":"Exists","effect":"NoExecute","tolerationSeconds":120},{"key":"node.kubernetes.io/not-ready","operator":"Exists","effect":"NoExecute","tolerationSeconds":120},{"key":"node.kubernetes.io/memory-pressure","operator":"Exists","effect":"NoSchedule"}],"priorityClassName":"system-cluster-critical","priority":2000000000,"enableServiceLinks":true,"preemptionPolicy":"PreemptLowerPriority"},"status":{"phase":"Running","conditions":[{"type":"PodReadyToStartContainers","status":"True","lastProbeTime":null,"lastTransitionTime":"2025-12-04T11:37:43Z"},{"type":"Initialized","status":"True","lastProbeTime":null,"lastTransitionTime":"2025-12-04T11:37:30Z"},{"type":"Ready","status":"True","lastProbeTime":null,"lastTransitionTime":"2025-12-04T11:55:48Z"},{"type":"ContainersReady","status":"True","lastProbeTime":null,"lastTransitionTime":"2025-12-04T11:55:48Z"},{"type":"PodScheduled","status":"True","lastProbeTime":null,"lastTransitionTime":"2025-12-04T11:37:30Z"}],"hostIP":"192.168.32.10","hostIPs":[{"ip":"192.168.32.10"}],"podIP":"10.128.0.8","podIPs":[{"ip":"10.128.0.8"}],"startTime":"2025-12-04T11:37:30Z","containerStatuses":[{"name":"openshift-controller-manager-operator","state":{"running":{"startedAt":"2025-12-04T11:55:47Z"}},"lastState":{"terminated":{"exitCode":0,"reason":"Completed","startedAt":"2025-12-04T11:43:54Z","finishedAt":"2025-12-04T11:55:46Z","containerID":"cri-o://a39ca33d8356cc3dbacc15a33d88f7f3eadcf0d16af1ac588f4e782ea1463d52"}},"ready":true,"restartCount":4,"image":"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8eabac819f289e29d75c7ab172d8124554849a47f0b00770928c3eb19a5a31c4","imageID":"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8eabac819f289e29d75c7ab172d8124554849a47f0b00770928c3eb19a5a31c4","containerID":"cri-o://6bc950db3ec1f7afbe92d39e38366383e0cef6eb9f745756096b42181a0164ac","started":true,"volumeMounts":[{"name":"config","mountPath":"/var/run/configmaps/config"},{"name":"serving-cert","mountPath":"/var/run/secrets/serving-cert"},{"name":"kube-api-access-k449s","mountPath":"/var/run/secrets/kubernetes.io/serviceaccount","readOnly":true,"recursiveReadOnly":"Disabled"}]}],"qosClass":"Burstable"}}config/pod/openshift-monitoring/prometheus-operator-admission-webhook-7c85c4dffd-xv2wn.json0000640000000000000000000000754215114274105030641 0ustar0000000000000000{"metadata":{"name":"prometheus-operator-admission-webhook-7c85c4dffd-xv2wn","generateName":"prometheus-operator-admission-webhook-7c85c4dffd-","namespace":"openshift-monitoring","uid":"ae0f644f-a31c-4b6c-b854-e01417aa448e","resourceVersion":"7168","creationTimestamp":"2025-12-04T11:38:33Z","labels":{"app.kubernetes.io/managed-by":"cluster-monitoring-operator","app.kubernetes.io/name":"prometheus-operator-admission-webhook","app.kubernetes.io/part-of":"openshift-monitoring","app.kubernetes.io/version":"0.78.1","pod-template-hash":"7c85c4dffd"},"annotations":{"kubectl.kubernetes.io/default-container":"prometheus-operator-admission-webhook","openshift.io/required-scc":"restricted-v2","openshift.io/scc":"restricted-v2","seccomp.security.alpha.kubernetes.io/pod":"runtime/default"},"ownerReferences":[{"apiVersion":"apps/v1","kind":"ReplicaSet","name":"prometheus-operator-admission-webhook-7c85c4dffd","uid":"7e206f35-bbf9-424e-9f7e-9f94e2838558","controller":true,"blockOwnerDeletion":true}]},"spec":{"volumes":[{"name":"tls-certificates","secret":{"secretName":"prometheus-operator-admission-webhook-tls","items":[{"key":"tls.crt","path":"tls.crt"},{"key":"tls.key","path":"tls.key"}],"defaultMode":420}}],"containers":[{"name":"prometheus-operator-admission-webhook","image":"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d2d169850894a59fb18012f5b1cde98a7e30fa5b86967c9d16e4cba5e88d9a8d","args":["--web.enable-tls=true","--web.cert-file=/etc/tls/private/tls.crt","--web.key-file=/etc/tls/private/tls.key","--web.tls-cipher-suites=TLS_AES_128_GCM_SHA256,TLS_AES_256_GCM_SHA384,TLS_CHACHA20_POLY1305_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305_SHA256,TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305_SHA256","--web.tls-min-version=VersionTLS12"],"ports":[{"name":"https","containerPort":8443,"protocol":"TCP"}],"resources":{"requests":{"cpu":"5m","memory":"30Mi"}},"volumeMounts":[{"name":"tls-certificates","readOnly":true,"mountPath":"/etc/tls/private"}],"livenessProbe":{"httpGet":{"path":"/healthz","port":"https","scheme":"HTTPS"},"timeoutSeconds":1,"periodSeconds":10,"successThreshold":1,"failureThreshold":3},"readinessProbe":{"httpGet":{"path":"/healthz","port":"https","scheme":"HTTPS"},"timeoutSeconds":1,"periodSeconds":10,"successThreshold":1,"failureThreshold":3},"terminationMessagePath":"/dev/termination-log","terminationMessagePolicy":"FallbackToLogsOnError","imagePullPolicy":"IfNotPresent","securityContext":{"capabilities":{"drop":["ALL"]},"runAsUser":1000460000,"runAsNonRoot":true,"allowPrivilegeEscalation":false}}],"restartPolicy":"Always","terminationGracePeriodSeconds":30,"dnsPolicy":"ClusterFirst","serviceAccountName":"prometheus-operator-admission-webhook","serviceAccount":"prometheus-operator-admission-webhook","automountServiceAccountToken":false,"securityContext":{"seLinuxOptions":{"level":"s0:c21,c20"},"fsGroup":1000460000,"seccompProfile":{"type":"RuntimeDefault"}},"schedulerName":"default-scheduler","tolerations":[{"key":"node.kubernetes.io/not-ready","operator":"Exists","effect":"NoExecute","tolerationSeconds":300},{"key":"node.kubernetes.io/unreachable","operator":"Exists","effect":"NoExecute","tolerationSeconds":300},{"key":"node.kubernetes.io/memory-pressure","operator":"Exists","effect":"NoSchedule"}],"priorityClassName":"system-cluster-critical","priority":2000000000,"enableServiceLinks":true,"preemptionPolicy":"PreemptLowerPriority"},"status":{"phase":"Pending","conditions":[{"type":"PodScheduled","status":"False","lastProbeTime":null,"lastTransitionTime":"2025-12-04T11:38:33Z","reason":"Unschedulable","message":"0/1 nodes are available: 1 node(s) had untolerated taint {node-role.kubernetes.io/master: }. preemption: 0/1 nodes are available: 1 Preemption is not helpful for scheduling."}],"qosClass":"Burstable"}}cluster-monitoring-operator_current.log0000640000000000000000000007113715114274105036310 0ustar0000000000000000config/pod/openshift-monitoring/logs/cluster-monitoring-operator-7ff994598c-p82nn2025-12-04T11:50:36.471970180Z I1204 11:50:36.471425 1 operator.go:647] Triggering an update due to a change in *v1.ClusterVersion/version 2025-12-04T11:50:36.488917132Z I1204 11:50:36.488816 1 operator.go:647] Triggering an update due to a change in *v1.ClusterVersion/version 2025-12-04T11:50:51.472356223Z I1204 11:50:51.470654 1 operator.go:647] Triggering an update due to a change in *v1.ClusterVersion/version 2025-12-04T11:50:53.503668077Z W1204 11:50:53.503614 1 tasks.go:73] task 2 of 2: UpdatingPrometheusOperator failed: reconciling Prometheus Operator Admission Webhook Deployment failed: updating Deployment object failed: waiting for DeploymentRollout of openshift-monitoring/prometheus-operator-admission-webhook: context deadline exceeded: got 1 unavailable replicas 2025-12-04T11:50:53.503749819Z I1204 11:50:53.503733 1 operator.go:894] 2 reconciliation(s) failed, 2 more attempt(s) will be made before reporting failures. 2025-12-04T11:50:53.503782040Z E1204 11:50:53.503772 1 operator.go:705] Syncing "openshift-monitoring/cluster-monitoring-config" failed 2025-12-04T11:50:53.503847352Z E1204 11:50:53.503832 1 operator.go:706] "Unhandled Error" err="sync \"openshift-monitoring/cluster-monitoring-config\" failed: cluster monitoring update failed (reason: UpdatingPrometheusOperatorFailed)" logger="UnhandledError" 2025-12-04T11:50:53.505231314Z W1204 11:50:53.505175 1 operator.go:993] No Cluster Monitoring ConfigMap was found. Using defaults. 2025-12-04T11:50:53.528866847Z I1204 11:50:53.528760 1 operator.go:855] Updating ClusterOperator status to InProgress. 2025-12-04T11:50:53.531260609Z I1204 11:50:53.531237 1 tasks.go:49] processing task group 1 of 3 2025-12-04T11:50:53.531515157Z I1204 11:50:53.531499 1 tasks.go:70] running task 2 of 2: UpdatingPrometheusOperator 2025-12-04T11:50:53.531559349Z I1204 11:50:53.531538 1 tasks.go:70] running task 1 of 2: UpdatingMetricsScrapingClientCA 2025-12-04T11:50:53.540523219Z I1204 11:50:53.540438 1 tasks.go:76] ran task 1 of 2: UpdatingMetricsScrapingClientCA 2025-12-04T11:53:32.866433002Z I1204 11:53:32.866322 1 operator.go:677] Triggering an update due to ConfigMap or Secret: openshift-monitoring/metrics-client-certs 2025-12-04T11:53:32.866433002Z I1204 11:53:32.866373 1 operator.go:677] Triggering an update due to ConfigMap or Secret: openshift-monitoring/federate-client-certs 2025-12-04T11:53:32.867470634Z I1204 11:53:32.867407 1 operator.go:677] Triggering an update due to ConfigMap or Secret: kube-system/extension-apiserver-authentication 2025-12-04T11:53:32.918216505Z I1204 11:53:32.918105 1 operator.go:677] Triggering an update due to ConfigMap or Secret: openshift-config-managed/kubelet-serving-ca 2025-12-04T11:53:33.150962967Z I1204 11:53:33.150743 1 operator.go:647] Triggering an update due to a change in *v1.APIServer/cluster 2025-12-04T11:53:33.150962967Z I1204 11:53:33.150821 1 operator.go:647] Triggering an update due to a change in *v1.ClusterVersion/version 2025-12-04T11:53:33.151104461Z I1204 11:53:33.151083 1 operator.go:647] Triggering an update due to a change in *v1.Infrastructure/cluster 2025-12-04T11:53:33.152044892Z I1204 11:53:33.152007 1 operator.go:647] Triggering an update due to a change in Console.config.openshift.io/cluster 2025-12-04T11:53:33.162086222Z I1204 11:53:33.162032 1 operator.go:606] ConfigMap to configure stack does not exist. Reconciling with default config every 5m0s. 2025-12-04T11:54:32.871394238Z E1204 11:54:32.871307 1 base_controller.go:277] "Unhandled Error" err="\"OpenShiftMonitoringTelemeterClientCertRequester\" controller failed to sync \"system:openshift:openshift-monitoring-62r5r\", err: unable to get secret \"openshift-monitoring/federate-client-certs\": the server was unable to return a response in the time allotted, but may still be processing the request (get secrets federate-client-certs)" logger="UnhandledError" 2025-12-04T11:54:32.871394238Z E1204 11:54:32.871375 1 base_controller.go:277] "Unhandled Error" err="\"OpenShiftMonitoringClientCertRequester\" controller failed to sync \"system:openshift:openshift-monitoring-62r5r\", err: unable to get secret \"openshift-monitoring/metrics-client-certs\": the server was unable to return a response in the time allotted, but may still be processing the request (get secrets metrics-client-certs)" logger="UnhandledError" 2025-12-04T11:54:33.171165700Z E1204 11:54:33.171045 1 relabel_controller.go:185] "Unhandled Error" err="Error syncing AlertRelabelConfig (secret/openshift-monitoring/alert-relabel-configs): the server was unable to return a response in the time allotted, but may still be processing the request (get secrets alert-relabel-configs)" logger="UnhandledError" 2025-12-04T11:55:48.503255130Z I1204 11:55:48.503135 1 operator.go:647] Triggering an update due to a change in *v1.ClusterVersion/version 2025-12-04T11:55:53.866393548Z W1204 11:55:53.866328 1 tasks.go:73] task 2 of 2: UpdatingPrometheusOperator failed: reconciling Prometheus Operator Admission Webhook Deployment failed: updating Deployment object failed: waiting for DeploymentRollout of openshift-monitoring/prometheus-operator-admission-webhook: context deadline exceeded: got 1 unavailable replicas 2025-12-04T11:55:53.866552493Z I1204 11:55:53.866533 1 operator.go:894] 3 reconciliation(s) failed, 1 more attempt(s) will be made before reporting failures. 2025-12-04T11:55:53.866607325Z E1204 11:55:53.866594 1 operator.go:705] Syncing "openshift-monitoring/cluster-monitoring-config" failed 2025-12-04T11:55:53.866652146Z E1204 11:55:53.866641 1 operator.go:706] "Unhandled Error" err="sync \"openshift-monitoring/cluster-monitoring-config\" failed: cluster monitoring update failed (reason: UpdatingPrometheusOperatorFailed)" logger="UnhandledError" 2025-12-04T11:55:53.867853342Z W1204 11:55:53.867830 1 operator.go:993] No Cluster Monitoring ConfigMap was found. Using defaults. 2025-12-04T11:56:16.469654822Z I1204 11:56:16.469573 1 operator.go:855] Updating ClusterOperator status to InProgress. 2025-12-04T11:56:16.476487130Z I1204 11:56:16.476408 1 tasks.go:49] processing task group 1 of 3 2025-12-04T11:56:16.476487130Z I1204 11:56:16.476471 1 tasks.go:70] running task 2 of 2: UpdatingPrometheusOperator 2025-12-04T11:56:16.476586493Z I1204 11:56:16.476525 1 tasks.go:70] running task 1 of 2: UpdatingMetricsScrapingClientCA 2025-12-04T11:56:26.263466053Z I1204 11:56:26.262222 1 tasks.go:76] ran task 1 of 2: UpdatingMetricsScrapingClientCA 2025-12-04T11:56:29.926467690Z I1204 11:56:29.924503 1 operator.go:647] Triggering an update due to a change in *v1.ClusterVersion/version 2025-12-04T11:56:29.940676002Z I1204 11:56:29.940621 1 operator.go:647] Triggering an update due to a change in *v1.ClusterVersion/version 2025-12-04T11:56:29.981779171Z I1204 11:56:29.981351 1 operator.go:647] Triggering an update due to a change in *v1.ClusterVersion/version 2025-12-04T11:56:29.994540178Z I1204 11:56:29.990172 1 operator.go:647] Triggering an update due to a change in *v1.ClusterVersion/version 2025-12-04T11:56:44.675506849Z I1204 11:56:44.674529 1 operator.go:647] Triggering an update due to a change in *v1.ClusterVersion/version 2025-12-04T11:56:44.689444622Z I1204 11:56:44.689369 1 operator.go:647] Triggering an update due to a change in *v1.ClusterVersion/version 2025-12-04T11:56:59.676174364Z I1204 11:56:59.675620 1 operator.go:647] Triggering an update due to a change in *v1.ClusterVersion/version 2025-12-04T11:58:33.162190523Z I1204 11:58:33.162071 1 operator.go:606] ConfigMap to configure stack does not exist. Reconciling with default config every 5m0s. 2025-12-04T11:59:33.613412731Z I1204 11:59:33.613289 1 operator.go:647] Triggering an update due to a change in *v1.ClusterVersion/version 2025-12-04T11:59:35.395274264Z I1204 11:59:35.395190 1 operator.go:647] Triggering an update due to a change in *v1.Infrastructure/cluster 2025-12-04T11:59:37.044389594Z I1204 11:59:37.044326 1 operator.go:647] Triggering an update due to a change in *v1.APIServer/cluster 2025-12-04T11:59:44.674001832Z I1204 11:59:44.673919 1 operator.go:647] Triggering an update due to a change in *v1.ClusterVersion/version 2025-12-04T11:59:46.314026293Z I1204 11:59:46.313930 1 reflector.go:368] Caches populated for *v1.ConfigMap from github.com/openshift/cluster-monitoring-operator/pkg/operator/operator.go:566 2025-12-04T11:59:46.677789254Z I1204 11:59:46.677696 1 reflector.go:368] Caches populated for *v1.FeatureGate from github.com/openshift/client-go/config/informers/externalversions/factory.go:125 2025-12-04T11:59:49.754658412Z I1204 11:59:49.753734 1 reflector.go:368] Caches populated for *v1.AlertRelabelConfig from github.com/openshift/cluster-monitoring-operator/pkg/alert/relabel_controller.go:121 2025-12-04T11:59:50.425835519Z I1204 11:59:50.425741 1 reflector.go:368] Caches populated for *v1.Secret from github.com/openshift/cluster-monitoring-operator/pkg/operator/operator.go:566 2025-12-04T11:59:50.426016795Z I1204 11:59:50.425970 1 operator.go:677] Triggering an update due to ConfigMap or Secret: openshift-monitoring/federate-client-certs 2025-12-04T11:59:50.426016795Z I1204 11:59:50.426006 1 operator.go:677] Triggering an update due to ConfigMap or Secret: openshift-monitoring/metrics-client-certs 2025-12-04T11:59:51.170651428Z I1204 11:59:51.170566 1 reflector.go:368] Caches populated for *v1.AlertingRule from github.com/openshift/cluster-monitoring-operator/pkg/alert/rule_controller.go:118 2025-12-04T11:59:53.748458879Z I1204 11:59:53.748309 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/apiserver/pkg/server/dynamiccertificates/configmap_cafile_content.go:209 2025-12-04T11:59:53.749026847Z I1204 11:59:53.748958 1 tlsconfig.go:181] "Loaded client CA" index=0 certName="client-ca::kube-system::extension-apiserver-authentication::client-ca-file,client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file" certDetail="\"admin-kubeconfig-signer\" [] issuer=\"\" (2025-12-04 11:20:50 +0000 UTC to 2035-12-02 11:20:50 +0000 UTC (now=2025-12-04 11:59:53.748899973 +0000 UTC))" 2025-12-04T11:59:53.749026847Z I1204 11:59:53.749012 1 tlsconfig.go:181] "Loaded client CA" index=1 certName="client-ca::kube-system::extension-apiserver-authentication::client-ca-file,client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file" certDetail="\"kubelet-signer\" [] issuer=\"\" (2025-12-04 11:26:53 +0000 UTC to 2025-12-05 11:26:53 +0000 UTC (now=2025-12-04 11:59:53.748996766 +0000 UTC))" 2025-12-04T11:59:53.749044678Z I1204 11:59:53.749028 1 tlsconfig.go:181] "Loaded client CA" index=2 certName="client-ca::kube-system::extension-apiserver-authentication::client-ca-file,client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file" certDetail="\"kube-control-plane-signer\" [] issuer=\"\" (2025-12-04 11:26:54 +0000 UTC to 2026-12-04 11:26:54 +0000 UTC (now=2025-12-04 11:59:53.749017977 +0000 UTC))" 2025-12-04T11:59:53.749054128Z I1204 11:59:53.749046 1 tlsconfig.go:181] "Loaded client CA" index=3 certName="client-ca::kube-system::extension-apiserver-authentication::client-ca-file,client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file" certDetail="\"kube-apiserver-to-kubelet-signer\" [] issuer=\"\" (2025-12-04 11:26:54 +0000 UTC to 2026-12-04 11:26:54 +0000 UTC (now=2025-12-04 11:59:53.749034577 +0000 UTC))" 2025-12-04T11:59:53.749108680Z I1204 11:59:53.749070 1 tlsconfig.go:181] "Loaded client CA" index=4 certName="client-ca::kube-system::extension-apiserver-authentication::client-ca-file,client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file" certDetail="\"kubelet-bootstrap-kubeconfig-signer\" [] issuer=\"\" (2025-12-04 11:26:52 +0000 UTC to 2035-12-02 11:26:52 +0000 UTC (now=2025-12-04 11:59:53.749052048 +0000 UTC))" 2025-12-04T11:59:53.749108680Z I1204 11:59:53.749095 1 tlsconfig.go:181] "Loaded client CA" index=5 certName="client-ca::kube-system::extension-apiserver-authentication::client-ca-file,client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file" certDetail="\"kube-csr-signer_@1764848267\" [] issuer=\"kubelet-signer\" (2025-12-04 11:37:46 +0000 UTC to 2025-12-05 11:26:53 +0000 UTC (now=2025-12-04 11:59:53.749083889 +0000 UTC))" 2025-12-04T11:59:53.749144781Z I1204 11:59:53.749121 1 tlsconfig.go:181] "Loaded client CA" index=6 certName="client-ca::kube-system::extension-apiserver-authentication::client-ca-file,client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file" certDetail="\"openshift-kube-apiserver-operator_node-system-admin-signer@1764848255\" [] issuer=\"\" (2025-12-04 11:37:34 +0000 UTC to 2026-12-04 11:37:35 +0000 UTC (now=2025-12-04 11:59:53.74910099 +0000 UTC))" 2025-12-04T11:59:53.749177512Z I1204 11:59:53.749144 1 tlsconfig.go:181] "Loaded client CA" index=7 certName="client-ca::kube-system::extension-apiserver-authentication::client-ca-file,client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file" certDetail="\"aggregator-signer\" [] issuer=\"\" (2025-12-04 11:26:52 +0000 UTC to 2025-12-05 11:26:52 +0000 UTC (now=2025-12-04 11:59:53.749133851 +0000 UTC))" 2025-12-04T11:59:53.749403119Z I1204 11:59:53.749368 1 tlsconfig.go:203] "Loaded serving cert" certName="serving-cert::/etc/tls/private/tls.crt::/etc/tls/private/tls.key" certDetail="\"*.cluster-monitoring-operator.openshift-monitoring.svc\" [serving] validServingFor=[*.cluster-monitoring-operator.openshift-monitoring.svc,*.cluster-monitoring-operator.openshift-monitoring.svc.cluster.local,cluster-monitoring-operator.openshift-monitoring.svc,cluster-monitoring-operator.openshift-monitoring.svc.cluster.local] issuer=\"openshift-service-serving-signer@1764848266\" (2025-12-04 11:38:00 +0000 UTC to 2027-12-04 11:38:01 +0000 UTC (now=2025-12-04 11:59:53.749336037 +0000 UTC))" 2025-12-04T11:59:53.749554974Z I1204 11:59:53.749520 1 named_certificates.go:53] "Loaded SNI cert" index=0 certName="self-signed loopback" certDetail="\"apiserver-loopback-client@1764848313\" [serving] validServingFor=[apiserver-loopback-client] issuer=\"apiserver-loopback-client-ca@1764848313\" (2025-12-04 10:38:32 +0000 UTC to 2026-12-04 10:38:32 +0000 UTC (now=2025-12-04 11:59:53.749503192 +0000 UTC))" 2025-12-04T11:59:55.535543718Z I1204 11:59:55.535444 1 reflector.go:368] Caches populated for *v1.ConfigMap from github.com/openshift/cluster-monitoring-operator/pkg/operator/operator.go:566 2025-12-04T11:59:55.536072345Z I1204 11:59:55.535962 1 operator.go:677] Triggering an update due to ConfigMap or Secret: openshift-config-managed/kubelet-serving-ca 2025-12-04T11:59:56.133260979Z I1204 11:59:56.133156 1 reflector.go:368] Caches populated for *v1.ConfigMap from github.com/openshift/cluster-monitoring-operator/pkg/operator/operator.go:566 2025-12-04T11:59:56.133346352Z I1204 11:59:56.133308 1 operator.go:677] Triggering an update due to ConfigMap or Secret: kube-system/extension-apiserver-authentication 2025-12-04T11:59:56.720155016Z I1204 11:59:56.720061 1 reflector.go:368] Caches populated for *v1.ConfigMap from github.com/openshift/cluster-monitoring-operator/pkg/operator/operator.go:566 2025-12-04T11:59:57.133870678Z I1204 11:59:57.132798 1 reflector.go:368] Caches populated for *v1.ClusterVersion from github.com/openshift/client-go/config/informers/externalversions/factory.go:125 2025-12-04T11:59:57.873889381Z I1204 11:59:57.872406 1 reflector.go:368] Caches populated for *v1.PrometheusRule from github.com/openshift/cluster-monitoring-operator/pkg/alert/rule_controller.go:117 2025-12-04T11:59:58.267592595Z I1204 11:59:58.267497 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/apiserver/pkg/server/dynamiccertificates/configmap_cafile_content.go:209 2025-12-04T11:59:58.683690374Z I1204 11:59:58.683575 1 reflector.go:368] Caches populated for *v1.Console from github.com/openshift/cluster-monitoring-operator/pkg/operator/operator.go:566 2025-12-04T11:59:58.683764846Z I1204 11:59:58.683735 1 operator.go:647] Triggering an update due to a change in Console.config.openshift.io/cluster 2025-12-04T11:59:59.681894507Z I1204 11:59:59.680754 1 operator.go:647] Triggering an update due to a change in *v1.ClusterVersion/version 2025-12-04T11:59:59.705918818Z I1204 11:59:59.705795 1 reflector.go:368] Caches populated for *v1.CertificateSigningRequest from k8s.io/client-go/informers/factory.go:160 2025-12-04T11:59:59.748348752Z I1204 11:59:59.748251 1 reflector.go:368] Caches populated for *v1.ConfigMap from github.com/openshift/cluster-monitoring-operator/pkg/operator/operator.go:563 2025-12-04T12:00:00.347682854Z I1204 12:00:00.347567 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/apiserver/pkg/authentication/request/headerrequest/requestheader_controller.go:175 2025-12-04T12:00:01.449121028Z I1204 12:00:01.449044 1 reflector.go:368] Caches populated for *v1.Secret from k8s.io/client-go/informers/factory.go:160 2025-12-04T12:00:03.714863365Z I1204 12:00:03.714759 1 reflector.go:368] Caches populated for *v1.Secret from github.com/openshift/cluster-monitoring-operator/pkg/alert/relabel_controller.go:122 2025-12-04T12:00:29.678168390Z I1204 12:00:29.677143 1 operator.go:647] Triggering an update due to a change in *v1.ClusterVersion/version 2025-12-04T12:00:59.693891378Z I1204 12:00:59.690548 1 operator.go:647] Triggering an update due to a change in *v1.ClusterVersion/version 2025-12-04T12:00:59.719528010Z I1204 12:00:59.715671 1 operator.go:647] Triggering an update due to a change in *v1.ClusterVersion/version 2025-12-04T12:01:14.677104042Z I1204 12:01:14.677009 1 operator.go:647] Triggering an update due to a change in *v1.ClusterVersion/version 2025-12-04T12:01:16.504909312Z W1204 12:01:16.504826 1 tasks.go:73] task 2 of 2: UpdatingPrometheusOperator failed: reconciling Prometheus Operator Admission Webhook Deployment failed: updating Deployment object failed: waiting for DeploymentRollout of openshift-monitoring/prometheus-operator-admission-webhook: context deadline exceeded: got 1 unavailable replicas 2025-12-04T12:01:16.504909312Z I1204 12:01:16.504879 1 operator.go:897] 4 reconciliations failed in a row, the threshold of 4 attempts has been reached, failures will be reported. 2025-12-04T12:01:16.557251158Z E1204 12:01:16.556766 1 operator.go:705] Syncing "openshift-monitoring/cluster-monitoring-config" failed 2025-12-04T12:01:16.557251158Z E1204 12:01:16.556848 1 operator.go:706] "Unhandled Error" err="sync \"openshift-monitoring/cluster-monitoring-config\" failed: cluster monitoring update failed (reason: UpdatingPrometheusOperatorFailed)" logger="UnhandledError" 2025-12-04T12:01:16.558762617Z W1204 12:01:16.558701 1 operator.go:993] No Cluster Monitoring ConfigMap was found. Using defaults. 2025-12-04T12:01:16.582014612Z I1204 12:01:16.581902 1 operator.go:855] Updating ClusterOperator status to InProgress. 2025-12-04T12:01:16.585553424Z I1204 12:01:16.585485 1 tasks.go:49] processing task group 1 of 3 2025-12-04T12:01:16.585553424Z I1204 12:01:16.585537 1 tasks.go:70] running task 2 of 2: UpdatingPrometheusOperator 2025-12-04T12:01:16.585582615Z I1204 12:01:16.585567 1 tasks.go:70] running task 1 of 2: UpdatingMetricsScrapingClientCA 2025-12-04T12:01:16.607477798Z I1204 12:01:16.607372 1 core.go:352] ConfigMap "openshift-monitoring/metrics-client-ca" changes: {"apiVersion":"v1","data":{"client-ca.crt":"-----BEGIN CERTIFICATE-----\nMIIDMDCCAhigAwIBAgIIR57fXfj+L88wDQYJKoZIhvcNAQELBQAwNjESMBAGA1UE\nCxMJb3BlbnNoaWZ0MSAwHgYDVQQDExdhZG1pbi1rdWJlY29uZmlnLXNpZ25lcjAe\nFw0yNTEyMDQxMTIwNTBaFw0zNTEyMDIxMTIwNTBaMDYxEjAQBgNVBAsTCW9wZW5z\naGlmdDEgMB4GA1UEAxMXYWRtaW4ta3ViZWNvbmZpZy1zaWduZXIwggEiMA0GCSqG\nSIb3DQEBAQUAA4IBDwAwggEKAoIBAQDSrufqceT/wc57Xx2xOm88aw9ZLpF9GEb7\npCYLFjoSMB26h3BDlRJgSMF9xz9G6xWyf9ipfHwSZ9/rO4c8XwkCseGfyod1t2GE\ndx5RuUp185sBL9Wxh8K2dBZJFGPLMrL9nMghsutMdEVy6SuQ2J2uw7i7BBGCkBn7\nO+AlrZ6sPPf1c95+U5TvK6BZZdSc4Wjy3iuatcq94fhEmtfmrb9S/c6+dwUP/IRU\nQp/pnMzvjy/AvLkeUuxP2q9w5PahbSJcAOMz/guOO/IonCAjvNGyHRu7pKWFxQUs\n0I1j2SsTY0KDYNImOJU+5JuHgdh21msVecKQOCEX1MnEtqVpsjMNAgMBAAGjQjBA\nMA4GA1UdDwEB/wQEAwICpDAPBgNVHRMBAf8EBTADAQH/MB0GA1UdDgQWBBSvTAMS\n7UvROzk7PzI1hTiW3YwdjDANBgkqhkiG9w0BAQsFAAOCAQEAyf33YKKi57Cy8O5b\n9whfa28ib43vkjygf1UYYFnScRMFCyxwhe/yOPTecudDIKA9IHisoo/BoYJSQRJp\nydXvHTNNurO3LqOI3pNOkeZ+z3c1N43ml2fSIGZ34NuKUyB8NAoqv0nTJMBAPivy\nvhrIXW4YRqdooWSCKrLxl10pmyMiRf/nvrOmpuP1qvPDLblwb3e1+YBw94M4Gss4\nl5E7if8oQaEHkxMyIsBroTUFD3FFBKXHwr/hAO0D/wSBc4EGj+Pf+5aJh4RLp29J\n32FCamAP6Yd5+7Gb1Avsem8+hMKRIDyQ7ZK8OpOSqFzcQgiJ4d5ESl/HzuHJXxOl\nOp/LJg==\n-----END CERTIFICATE-----\n-----BEGIN CERTIFICATE-----\nMIIDHjCCAgagAwIBAgIIY7bo6FLvZzowDQYJKoZIhvcNAQELBQAwLTESMBAGA1UE\nCxMJb3BlbnNoaWZ0MRcwFQYDVQQDEw5rdWJlbGV0LXNpZ25lcjAeFw0yNTEyMDQx\nMTI2NTNaFw0yNTEyMDUxMTI2NTNaMC0xEjAQBgNVBAsTCW9wZW5zaGlmdDEXMBUG\nA1UEAxMOa3ViZWxldC1zaWduZXIwggEiMA0GCSqGSIb3DQEBAQUAA4IBDwAwggEK\nAoIBAQC5JvPPN2kWvgCPCS2q/i0DIvK9NQoE/x3GPW2zVreKMoEYNuhuHJR2QAr/\ngX8rbr+ZPQD4m71Gbxp5TQiyABWY2AGP3cYGvdQfNQ/6le2fvhzIl39BkFnTqBnj\nRRd1fST0AklMAZ1crVtOf8R2VjXM2CVyLB5bW89+HWNkIK4QAcxbyvWlxMLlpZp/\n66lsPh24Jf2GLyxRvWMEMvzKf8dc0bJFZxLAbDQzKqSDJsJCQoXtBSo7h+o7mPej\nfDIXnyOkdQKDuyE2j9VUl/o/ucxcavDoWu6PrNzva8tLCLMxgxXxaIfVf8SxCBk/\nMariw0NZaWlkeDxAOiZxqG7xsqBLAgMBAAGjQjBAMA4GA1UdDwEB/wQEAwICpDAP\nBgNVHRMBAf8EBTADAQH/MB0GA1UdDgQWBBSUrxu9k8lnfbP1KBQOjhPBk+C+VTAN\nBgkqhkiG9w0BAQsFAAOCAQEARiQc+tkQ28HOsVtMhgYmbpZsAB7KS4qEWtjMBbib\nOaCsJc0QliQ13Zr0MBIbNOKhwWDfzWyDYcR2XZaQd70P9d0F73xNuIGuXMD+fL7g\n+AXxiT5qlDrn/qkZ67mJ1plYtv7NiiwuSdxwz/NrGJ08y843WA+jhwTmTgWqq0Me\nfEFRvc5t2pWPV1BwyhrmQ6h55RfItH4JQ85Yl6L6M/5hrPvumyfmo2INpHKlXJdX\nsxvBb/iHXkEz+lEvtKgbxEb27yk74K5VbjxRO07E9pYuuMBQFYYpSsIl//nYjTBh\nhFUDv4ru4PO/xH2KGjhLA0PwcEz7ThI4XvhPN9wwUhW9Yw==\n-----END CERTIFICATE-----\n-----BEGIN CERTIFICATE-----\nMIIDNDCCAhygAwIBAgIIWjcloEQrA/AwDQYJKoZIhvcNAQELBQAwODESMBAGA1UE\nCxMJb3BlbnNoaWZ0MSIwIAYDVQQDExlrdWJlLWNvbnRyb2wtcGxhbmUtc2lnbmVy\nMB4XDTI1MTIwNDExMjY1NFoXDTI2MTIwNDExMjY1NFowODESMBAGA1UECxMJb3Bl\nbnNoaWZ0MSIwIAYDVQQDExlrdWJlLWNvbnRyb2wtcGxhbmUtc2lnbmVyMIIBIjAN\nBgkqhkiG9w0BAQEFAAOCAQ8AMIIBCgKCAQEAtTkV1bhmxNqMYJPXIuANuUzfQvr9\nsGpavJRCL+vC1N47w+admLPAfZ2CCGjEkgypHntoyTlFjrBcthw0qsYgTwGb4vt2\nWdP6mkJoQ/T2sa8kFlZSjtgCzME+Jvck4xTGty0QO9b26hK39f/n1pOIknEXli/A\nGnz7H57jGjmu8hhuHEKwqc9PKxD0AMwE3x7B7vvLeJfWR8NBmF2Rzml61RNUxqC7\nUCPM0g6nKMPbCGfjG/TiLd0r5hFHMrdmd+uQLVzfkIgeWG9bu05pZoB8hQ/6tcJv\ncZmGlj2nBN0d653aneRorqPx51zXznI2aPndGbiwCEEwK5fgkRMb1ki6iQIDAQAB\no0IwQDAOBgNVHQ8BAf8EBAMCAqQwDwYDVR0TAQH/BAUwAwEB/zAdBgNVHQ4EFgQU\nm9GWA3AwGvEpZ0vpIe8y4n7qO5EwDQYJKoZIhvcNAQELBQADggEBAAOc2JG9htVv\nfi4dPce2pfTUCdgw5N8fhbKnpJrcER+JNg9AyFwBSRQ3Y2iLUaBw0RhwwPSkO90F\nC6Ss+OK5gyTHwB3lCUleczgeI4/mPCcoXjTvVg7IivZ4VR61RvAga/yptMEOvmSf\n5FdQkvpmzX7rPu+kFvw6tZy3CN6NTZo+7Msb9aXhWyPd83iOpVoMJ0oyp58a2gGU\nlSk3vZH1DyBwMbTO4xknaVSYAHNbCThjw6Ef6GXW1tQLhkzKy5Va756DpI/C4X6j\ny37fwv90LYEQzmAXo7LNv4EycYqfsu1hbxbBKTk87/YGEoULMaVm3yaqD0t2/E+m\nJ1rY73wX7kM=\n-----END CERTIFICATE-----\n-----BEGIN CERTIFICATE-----\nMIIDQjCCAiqgAwIBAgIIWvYp+AqidIwwDQYJKoZIhvcNAQELBQAwPzESMBAGA1UE\nCxMJb3BlbnNoaWZ0MSkwJwYDVQQDEyBrdWJlLWFwaXNlcnZlci10by1rdWJlbGV0\nLXNpZ25lcjAeFw0yNTEyMDQxMTI2NTRaFw0yNjEyMDQxMTI2NTRaMD8xEjAQBgNV\nBAsTCW9wZW5zaGlmdDEpMCcGA1UEAxMga3ViZS1hcGlzZXJ2ZXItdG8ta3ViZWxl\ndC1zaWduZXIwggEiMA0GCSqGSIb3DQEBAQUAA4IBDwAwggEKAoIBAQDd9Httqr0I\nnlCAPulOh9BdTL9+anXAHvsrODsI/Cru/lOqKyoZM4T+j4UXb1BEEHQvnHwQi1n6\nGlusUICm2BvFH7+Jvd+wzyeA/8rOq0dOyssOEZny7YeB56iOrik2V64fBl72fF6b\n5rBJ48TFzPGPb8p3vtTVzj05q9FT8c6rjNWu9MeUOv2jtVXeKxLx9LJ/RL83y9Kc\n/Cs9CQgxp5BNjYPq+HQ61As14XosxGK/ikLUUiMYU1UEcoH+hYDo8Jbg97Fu8UlK\n0JG5tjMGIUEgLXokOVzumVuN7Owe8zZsFWuU8GURg1o5D3psgkpC9x6RCc45enom\nl8TRq29v4C0XAgMBAAGjQjBAMA4GA1UdDwEB/wQEAwICpDAPBgNVHRMBAf8EBTAD\nAQH/MB0GA1UdDgQWBBSfVpoYxPT4568waylKxqqEsqWjFzANBgkqhkiG9w0BAQsF\nAAOCAQEAAmzrtFlkEbiSUjIKedtAAsGOR91egUdi0gd0layBaTQ460D+BDPWwo9H\n/HfKd0MVHytrsL5A2vONOKI7ISLySiliFOi4tzJQ7iRhgfKcWMFMix6VONHCYxbm\njAlpUlzNNZR3+OnvmP5Tn2Uwm6srASUaB8BsyqyTCV4QTcvU5StpJTe769374Xyr\naJvLGYZSvFNV7j5DKJ2sFzDbGGBK8OEgyhbpkd4ooEapt32QanmbL29UtpxynlsE\nfqMdtWbQcG1BfS5o12JJGa1gSOLNKkRV8s1AFkPF588I0ZunNTc4hlk8J6DurE3l\nSCburJkTOdEE8yTMQ13JnUsu8dc17Q==\n-----END CERTIFICATE-----\n-----BEGIN CERTIFICATE-----\nMIIDSDCCAjCgAwIBAgIII2m4dI8wxPYwDQYJKoZIhvcNAQELBQAwQjESMBAGA1UE\nCxMJb3BlbnNoaWZ0MSwwKgYDVQQDEyNrdWJlbGV0LWJvb3RzdHJhcC1rdWJlY29u\nZmlnLXNpZ25lcjAeFw0yNTEyMDQxMTI2NTJaFw0zNTEyMDIxMTI2NTJaMEIxEjAQ\nBgNVBAsTCW9wZW5zaGlmdDEsMCoGA1UEAxMja3ViZWxldC1ib290c3RyYXAta3Vi\nZWNvbmZpZy1zaWduZXIwggEiMA0GCSqGSIb3DQEBAQUAA4IBDwAwggEKAoIBAQC6\nvHB5gCRjWFg9j0y1aMNqwo8MyZ0a3qITeKv0cUpJm7Y3mJizlPT6gryBJEFlmJzD\nwowS/QxqoOG5Ci0Zo2hB/KKp5wTM+jqGaIraiYcKTpS1rrE9UL0Z79VVr81/q93w\nZRB1ChKlNznsmdEAK5KfRGbm86EtGyfh4xjW8oC8oT9QNdFtV+QMevBNX5dt0T2F\nnLwLzBtfh2ZSSFTl5Fi2UgggDsAypuPNiV4O+OXpxKcylDOV0Hrd13CgJLWcCfNh\ndwwiWHpI9GKrcHJlj/qY4b3tT7+NMoAzXawTEDc69ri/RF1jaZ2ug69clUgILzEN\n6EiiI7hoJVK3+JALeOUHAgMBAAGjQjBAMA4GA1UdDwEB/wQEAwICpDAPBgNVHRMB\nAf8EBTADAQH/MB0GA1UdDgQWBBRJgC1bqOvWbPKTGZ3D6QOfENyXfjANBgkqhkiG\n9w0BAQsFAAOCAQEARZqzTpbBx57UwqMY2MSyv9ZWqVYdgtVC6tvLT10udOXX4KTY\nQ2aSCT2q4Q6SnBWA/M50CLio/Z6l6vNqDBDP5grvMgSAzWGqwXVvMKN/KLcMgsDt\nbxwgYNG3lewtMsqYBWT5vNu65PhuWbRKiDdRC72EGOZKpyFED4sSqns53zo3MJmD\nZ7eKYRXa/mPLRoR+2LqaSFJeox32cAESPJUd5Xrpm1ohshldN/NiyRu8UPPE9WUf\nos71XR4HQLFHQcrQk1X+W1FJoAi4bB9WqP+fDNeg+aIq3C94GrRc3x4luEWR2H9J\nFNc4fF0PxjDX5t60S2eD1V1SSSPsrkDcma7fTw==\n-----END CERTIFICATE-----\n-----BEGIN CERTIFICATE-----\nMIIDODCCAiCgAwIBAgIITl0eyVVz4lswDQYJKoZIhvcNAQELBQAwLTESMBAGA1UE\nCxMJb3BlbnNoaWZ0MRcwFQYDVQQDEw5rdWJlbGV0LXNpZ25lcjAeFw0yNTEyMDQx\nMTM3NDZaFw0yNTEyMDUxMTI2NTNaMCYxJDAiBgNVBAMMG2t1YmUtY3NyLXNpZ25l\ncl9AMTc2NDg0ODI2NzCCASIwDQYJKoZIhvcNAQEBBQADggEPADCCAQoCggEBAOe9\ndKj9gAtNBNXtWsItk798raGahjy0rxPud3WIEKR08rBQ0y6pVWeYVh0kPeXtv+KH\now1ffRiVjsStNFc0s8tL47zt2cvP5E05EKNuQwdowrxiYPKJ3/ETNJKc8TKd3KVS\njwkVXBO8Yh/I8dUC5GbSB5o4eqrAAN4t3fdF+vZxEB2s3aqQpZLi8OaRd0guRSqV\nIaxpxMfs+tYFSQIizJDz6P33vK4+eEAfyHrAu9bU/5BoK+2vZr6BmsXeqtulLmZN\nbSsRfMOa18XBL/Dya69/k7PkicNQvPHEHbGIiPp5vyb9NrDKTuPbxiOIT1g4ij8Q\nCb4tOJ7R1Wq+NrzGDMsCAwEAAaNjMGEwDgYDVR0PAQH/BAQDAgKkMA8GA1UdEwEB\n/wQFMAMBAf8wHQYDVR0OBBYEFIbIfmkw7ruiJvCd/QefvqR8ppCGMB8GA1UdIwQY\nMBaAFJSvG72TyWd9s/UoFA6OE8GT4L5VMA0GCSqGSIb3DQEBCwUAA4IBAQCyNJmN\n8JEp8dWBBxKmypfGKXub9XICKS/C8X9lHku6Pw1S+J5LuDg5NIX419/3W+CBbSYn\nP5FkXAvwOiz/QV/7GRCfPHvi1pDCi+2ip12l/6njc6vjiYUAFAp+qAskDw3MYD71\nWSUk+qmyqkkkNYMillq3AQwm4ik5DDrjW4zjuxDgJ2dGiuWt5nw/lu7LXTuW4gmt\nFE8x4FyGaZgFaeTYSI3f7LIFvSpapiXgyQYire942hNTBUwTDDx88U4rxxqWhOR/\nOjXaxS0uZ1GpBvsuvoLQcmS27j3g1cinMdTdlQVr5z1M6V69TXmTcu3wSEg2XVgK\n3obNxRtOon/c0Bdi\n-----END CERTIFICATE-----\n-----BEGIN CERTIFICATE-----\nMIIDhTCCAm2gAwIBAgIIe3WqU1YhwLAwDQYJKoZIhvcNAQELBQAwUDFOMEwGA1UE\nAwxFb3BlbnNoaWZ0LWt1YmUtYXBpc2VydmVyLW9wZXJhdG9yX25vZGUtc3lzdGVt\nLWFkbWluLXNpZ25lckAxNzY0ODQ4MjU1MB4XDTI1MTIwNDExMzczNFoXDTI2MTIw\nNDExMzczNVowUDFOMEwGA1UEAwxFb3BlbnNoaWZ0LWt1YmUtYXBpc2VydmVyLW9w\nZXJhdG9yX25vZGUtc3lzdGVtLWFkbWluLXNpZ25lckAxNzY0ODQ4MjU1MIIBIjAN\nBgkqhkiG9w0BAQEFAAOCAQ8AMIIBCgKCAQEAzRK8/uYvwocpSCYjl+ZxXmJDBACK\nLhqPRVUATIMkGq/+jx9ylZgZopwc4AQiVwbLRv71pM96PVaF2bEhznQ/RXN6ECBd\nb/fIdJw2yg7wK7g+2q534hLfAm3C85jh58j2iU7eQ57iKRLeiaPQzUsO9kcDGAIt\n8jRel8d+e2fuk8911rm+O3fY5lNQy1CQ+E0ISDWhkaPs/NM2XK1hYn7ehEiVG92H\nRKgbock2Axfie6h9QooF4FLve5WRK7dGnHg+c95OwORhenPoI4mN4oBH8V6hxZlK\n9RIGXNsRlqOs3ha0TVK6y9Fx4T/iaRaxuGp8RTcQ2OVwof3rNO6mDleqUQIDAQAB\no2MwYTAOBgNVHQ8BAf8EBAMCAqQwDwYDVR0TAQH/BAUwAwEB/zAdBgNVHQ4EFgQU\nabKBDmu7hgcXkpplt2Ra9hIYLukwHwYDVR0jBBgwFoAUabKBDmu7hgcXkpplt2Ra\n9hIYLukwDQYJKoZIhvcNAQELBQADggEBACJXlIxfytsFL3jOxigPo7qnGBYeD4pg\nObBd9VhBojKujv0CsNmzvAq1ffkP0NpdSTfKWZs8M9BR230I5vGDylH8hRhy4QB3\n7mffxYjvVWrYofXx1zupR39DnWjUvy3PeYuFDnbfp4mkAA98OxKPMb6fb/yHqlQm\nLiBj4CdsemWd47dt2nRv40j4C6AfC2wa/3UH+0wS7pgn/stiAttmPT1EquwHyCaG\nw5rOJtEl96iA9w18EMZQkSabUcPACShfT4SkeEPlfIGmwHyMvdsyXsw5NaiULnlU\np6+lKK58gxqFBLq4icPMrOd141yZ34Q7CnxPZlvoa7HLcdUnSBytttc=\n-----END CERTIFICATE-----\n"},"kind":"ConfigMap","metadata":{"creationTimestamp":null,"managedFields":null,"resourceVersion":null,"uid":null}} 2025-12-04T12:01:16.607616762Z I1204 12:01:16.607576 1 event.go:377] Event(v1.ObjectReference{Kind:"Deployment", Namespace:"openshift-monitoring", Name:"cluster-monitoring-operator", UID:"c5bee1b0-c6f3-4003-b26e-809b1061fe43", APIVersion:"apps/v1", ResourceVersion:"", FieldPath:""}): type: 'Normal' reason: 'ConfigMapUpdated' Updated ConfigMap/metrics-client-ca -n openshift-monitoring: 2025-12-04T12:01:16.607616762Z cause by changes in data.client-ca.crt 2025-12-04T12:01:16.610485763Z I1204 12:01:16.610425 1 tasks.go:76] ran task 1 of 2: UpdatingMetricsScrapingClientCA kube-rbac-proxy_current.log0000640000000000000000000000202415114274105035157 0ustar0000000000000000config/pod/openshift-machine-config-operator/logs/machine-config-operator-dc5d7666f-p2cmn2025-12-04T11:39:22.059098694Z W1204 11:39:22.058844 1 deprecated.go:66] 2025-12-04T11:39:22.059098694Z ==== Removed Flag Warning ====================== 2025-12-04T11:39:22.059098694Z 2025-12-04T11:39:22.059098694Z logtostderr is removed in the k8s upstream and has no effect any more. 2025-12-04T11:39:22.059098694Z 2025-12-04T11:39:22.059098694Z =============================================== 2025-12-04T11:39:22.059098694Z 2025-12-04T11:39:22.059098694Z I1204 11:39:22.059037 1 kube-rbac-proxy.go:530] Reading config file: /etc/kube-rbac-proxy/config-file.yaml 2025-12-04T11:39:22.059721693Z I1204 11:39:22.059689 1 kube-rbac-proxy.go:233] Valid token audiences: 2025-12-04T11:39:22.061331415Z I1204 11:39:22.061255 1 kube-rbac-proxy.go:347] Reading certificate files 2025-12-04T11:39:22.061645015Z I1204 11:39:22.061596 1 kube-rbac-proxy.go:395] Starting TCP socket on 0.0.0.0:9001 2025-12-04T11:39:22.062223244Z I1204 11:39:22.062191 1 kube-rbac-proxy.go:402] Listening securely on 0.0.0.0:9001 machine-config-operator_current.log0000640000000000000000000001620315114274105036651 0ustar0000000000000000config/pod/openshift-machine-config-operator/logs/machine-config-operator-dc5d7666f-p2cmn2025-12-04T11:53:33.396516647Z I1204 11:53:33.396400 1 start.go:52] Version: 4.18.29 (Raw: 21d5ae96, Hash: bb2aa85171d93b2df952ed802a8cb200164e666f) 2025-12-04T11:53:33.396699473Z I1204 11:53:33.396676 1 leaderelection.go:121] The leader election gives 4 retries and allows for 30s of clock skew. The kube-apiserver downtime tolerance is 78s. Worst non-graceful lease acquisition is 2m43s. Worst graceful lease acquisition is {26s}. 2025-12-04T11:53:33.396827477Z I1204 11:53:33.396781 1 metrics.go:92] Registering Prometheus metrics 2025-12-04T11:53:33.396931681Z I1204 11:53:33.396910 1 metrics.go:99] Starting metrics listener on 127.0.0.1:8797 2025-12-04T11:54:33.404075157Z W1204 11:54:33.403968 1 helpers.go:77] unable to get cluster infrastructure status, using HA cluster values for leader election: the server was unable to return a response in the time allotted, but may still be processing the request (get infrastructures.config.openshift.io cluster) 2025-12-04T11:54:33.405024977Z I1204 11:54:33.404954 1 leaderelection.go:254] attempting to acquire leader lease openshift-machine-config-operator/machine-config... 2025-12-04T11:55:33.409693747Z E1204 11:55:33.409602 1 leaderelection.go:436] error retrieving resource lock openshift-machine-config-operator/machine-config: the server was unable to return a response in the time allotted, but may still be processing the request (get leases.coordination.k8s.io machine-config) 2025-12-04T11:59:07.407750942Z I1204 11:59:07.407658 1 leaderelection.go:268] successfully acquired lease openshift-machine-config-operator/machine-config 2025-12-04T11:59:08.761087363Z I1204 11:59:08.760840 1 simple_featuregate_reader.go:171] Starting feature-gate-detector 2025-12-04T11:59:08.776120439Z I1204 11:59:08.776014 1 event.go:377] Event(v1.ObjectReference{Kind:"Node", Namespace:"openshift-machine-config-operator", Name:"master-0", UID:"9618c86a-8a87-4c00-8dd2-9747684d64b3", APIVersion:"v1", ResourceVersion:"", FieldPath:""}): type: 'Normal' reason: 'FeatureGatesInitialized' FeatureGates updated to featuregates.Features{Enabled:[]v1.FeatureGateName{"AWSEFSDriverVolumeMetrics", "AdminNetworkPolicy", "AlibabaPlatform", "AzureWorkloadIdentity", "BareMetalLoadBalancer", "BuildCSIVolumes", "ChunkSizeMiB", "CloudDualStackNodeIPs", "DisableKubeletCloudCredentialProviders", "GCPLabelsTags", "HardwareSpeed", "IngressControllerLBSubnetsAWS", "KMSv1", "ManagedBootImages", "ManagedBootImagesAWS", "MultiArchInstallAWS", "MultiArchInstallGCP", "NetworkDiagnosticsConfig", "NetworkLiveMigration", "NetworkSegmentation", "NewOLM", "NodeDisruptionPolicy", "OnClusterBuild", "PersistentIPsForVirtualization", "PrivateHostedZoneAWS", "SetEIPForNLBIngressController", "VSphereControlPlaneMachineSet", "VSphereDriverConfiguration", "VSphereMultiVCenters", "VSphereStaticIPs", "ValidatingAdmissionPolicy"}, Disabled:[]v1.FeatureGateName{"AWSClusterHostedDNS", "AdditionalRoutingCapabilities", "AutomatedEtcdBackup", "BootcNodeManagement", "CSIDriverSharedResource", "ClusterAPIInstall", "ClusterAPIInstallIBMCloud", "ClusterMonitoringConfig", "ConsolePluginContentSecurityPolicy", "DNSNameResolver", "DynamicResourceAllocation", "EtcdBackendQuota", "EventedPLEG", "Example", "ExternalOIDC", "ExternalOIDCWithUIDAndExtraClaimMappings", "GCPClusterHostedDNS", "GatewayAPI", "ImageStreamImportMode", "IngressControllerDynamicConfigurationManager", "InsightsConfig", "InsightsConfigAPI", "InsightsOnDemandDataGather", "InsightsRuntimeExtractor", "MachineAPIMigration", "MachineAPIOperatorDisableMachineHealthCheckController", "MachineAPIProviderOpenStack", "MachineConfigNodes", "MaxUnavailableStatefulSet", "MetricsCollectionProfiles", "MinimumKubeletVersion", "MixedCPUsAllocation", "MultiArchInstallAzure", "NodeSwap", "NutanixMultiSubnets", "OVNObservability", "OpenShiftPodSecurityAdmission", "PinnedImages", "PlatformOperators", "ProcMountType", "RouteAdvertisements", "RouteExternalCertificate", "ServiceAccountTokenNodeBinding", "SignatureStores", "SigstoreImageVerification", "StreamingCollectionEncodingToJSON", "StreamingCollectionEncodingToProtobuf", "TranslateStreamCloseWebsocketRequests", "UpgradeStatus", "UserNamespacesPodSecurityStandards", "UserNamespacesSupport", "VSphereMultiNetworks", "VolumeAttributesClass", "VolumeGroupSnapshot"}} 2025-12-04T11:59:08.776120439Z I1204 11:59:08.776000 1 start.go:132] FeatureGates initialized: knownFeatureGates=[AWSEFSDriverVolumeMetrics AdminNetworkPolicy AlibabaPlatform AzureWorkloadIdentity BareMetalLoadBalancer BuildCSIVolumes ChunkSizeMiB CloudDualStackNodeIPs DisableKubeletCloudCredentialProviders GCPLabelsTags HardwareSpeed IngressControllerLBSubnetsAWS KMSv1 ManagedBootImages ManagedBootImagesAWS MultiArchInstallAWS MultiArchInstallGCP NetworkDiagnosticsConfig NetworkLiveMigration NetworkSegmentation NewOLM NodeDisruptionPolicy OnClusterBuild PersistentIPsForVirtualization PrivateHostedZoneAWS SetEIPForNLBIngressController VSphereControlPlaneMachineSet VSphereDriverConfiguration VSphereMultiVCenters VSphereStaticIPs ValidatingAdmissionPolicy AWSClusterHostedDNS AdditionalRoutingCapabilities AutomatedEtcdBackup BootcNodeManagement CSIDriverSharedResource ClusterAPIInstall ClusterAPIInstallIBMCloud ClusterMonitoringConfig ConsolePluginContentSecurityPolicy DNSNameResolver DynamicResourceAllocation EtcdBackendQuota EventedPLEG Example ExternalOIDC ExternalOIDCWithUIDAndExtraClaimMappings GCPClusterHostedDNS GatewayAPI ImageStreamImportMode IngressControllerDynamicConfigurationManager InsightsConfig InsightsConfigAPI InsightsOnDemandDataGather InsightsRuntimeExtractor MachineAPIMigration MachineAPIOperatorDisableMachineHealthCheckController MachineAPIProviderOpenStack MachineConfigNodes MaxUnavailableStatefulSet MetricsCollectionProfiles MinimumKubeletVersion MixedCPUsAllocation MultiArchInstallAzure NodeSwap NutanixMultiSubnets OVNObservability OpenShiftPodSecurityAdmission PinnedImages PlatformOperators ProcMountType RouteAdvertisements RouteExternalCertificate ServiceAccountTokenNodeBinding SignatureStores SigstoreImageVerification StreamingCollectionEncodingToJSON StreamingCollectionEncodingToProtobuf TranslateStreamCloseWebsocketRequests UpgradeStatus UserNamespacesPodSecurityStandards UserNamespacesSupport VSphereMultiNetworks VolumeAttributesClass VolumeGroupSnapshot] 2025-12-04T11:59:08.882568110Z I1204 11:59:08.882471 1 operator.go:436] Change observed to kube-apiserver-server-ca 2025-12-04T11:59:09.133326023Z I1204 11:59:09.133243 1 operator.go:436] Change observed to kube-apiserver-server-ca 2025-12-04T11:59:10.094372330Z I1204 11:59:10.094300 1 operator.go:389] On-cluster layering featuregate enabled, starting MachineOSConfig informer 2025-12-04T11:59:10.194963176Z I1204 11:59:10.194879 1 operator.go:416] Starting MachineConfigOperator 2025-12-04T11:59:10.195497372Z E1204 11:59:10.195477 1 sync.go:2172] exiting image registry secrets fetch - machine-os-puller service account does not exist yet. 2025-12-04T11:59:50.263869949Z I1204 11:59:50.263745 1 operator.go:436] Change observed to kube-apiserver-server-ca 2025-12-04T11:59:56.144312259Z I1204 11:59:56.144227 1 operator.go:436] Change observed to kube-apiserver-server-ca machine-config-operator_previous.log0000640000000000000000000002572315114274105037052 0ustar0000000000000000config/pod/openshift-machine-config-operator/logs/machine-config-operator-dc5d7666f-p2cmn2025-12-04T11:39:21.602925446Z I1204 11:39:21.602613 1 start.go:52] Version: 4.18.29 (Raw: 21d5ae96, Hash: bb2aa85171d93b2df952ed802a8cb200164e666f) 2025-12-04T11:39:21.603086861Z I1204 11:39:21.602906 1 leaderelection.go:121] The leader election gives 4 retries and allows for 30s of clock skew. The kube-apiserver downtime tolerance is 78s. Worst non-graceful lease acquisition is 2m43s. Worst graceful lease acquisition is {26s}. 2025-12-04T11:39:21.603165284Z I1204 11:39:21.603065 1 metrics.go:92] Registering Prometheus metrics 2025-12-04T11:39:21.603235515Z I1204 11:39:21.603207 1 metrics.go:99] Starting metrics listener on 127.0.0.1:8797 2025-12-04T11:40:21.616941016Z W1204 11:40:21.616838 1 helpers.go:77] unable to get cluster infrastructure status, using HA cluster values for leader election: the server was unable to return a response in the time allotted, but may still be processing the request (get infrastructures.config.openshift.io cluster) 2025-12-04T11:40:21.618747145Z I1204 11:40:21.618686 1 leaderelection.go:254] attempting to acquire leader lease openshift-machine-config-operator/machine-config... 2025-12-04T11:41:21.623927202Z E1204 11:41:21.623248 1 leaderelection.go:436] error retrieving resource lock openshift-machine-config-operator/machine-config: the server was unable to return a response in the time allotted, but may still be processing the request (get leases.coordination.k8s.io machine-config) 2025-12-04T11:43:16.705427657Z E1204 11:43:16.705329 1 leaderelection.go:436] error retrieving resource lock openshift-machine-config-operator/machine-config: the server was unable to return a response in the time allotted, but may still be processing the request (get leases.coordination.k8s.io machine-config) 2025-12-04T11:44:15.047484750Z I1204 11:44:15.047174 1 leaderelection.go:268] successfully acquired lease openshift-machine-config-operator/machine-config 2025-12-04T11:45:10.764900702Z I1204 11:45:10.764735 1 simple_featuregate_reader.go:171] Starting feature-gate-detector 2025-12-04T11:45:10.796054899Z I1204 11:45:10.795951 1 start.go:132] FeatureGates initialized: knownFeatureGates=[AWSEFSDriverVolumeMetrics AdminNetworkPolicy AlibabaPlatform AzureWorkloadIdentity BareMetalLoadBalancer BuildCSIVolumes ChunkSizeMiB CloudDualStackNodeIPs DisableKubeletCloudCredentialProviders GCPLabelsTags HardwareSpeed IngressControllerLBSubnetsAWS KMSv1 ManagedBootImages ManagedBootImagesAWS MultiArchInstallAWS MultiArchInstallGCP NetworkDiagnosticsConfig NetworkLiveMigration NetworkSegmentation NewOLM NodeDisruptionPolicy OnClusterBuild PersistentIPsForVirtualization PrivateHostedZoneAWS SetEIPForNLBIngressController VSphereControlPlaneMachineSet VSphereDriverConfiguration VSphereMultiVCenters VSphereStaticIPs ValidatingAdmissionPolicy AWSClusterHostedDNS AdditionalRoutingCapabilities AutomatedEtcdBackup BootcNodeManagement CSIDriverSharedResource ClusterAPIInstall ClusterAPIInstallIBMCloud ClusterMonitoringConfig ConsolePluginContentSecurityPolicy DNSNameResolver DynamicResourceAllocation EtcdBackendQuota EventedPLEG Example ExternalOIDC ExternalOIDCWithUIDAndExtraClaimMappings GCPClusterHostedDNS GatewayAPI ImageStreamImportMode IngressControllerDynamicConfigurationManager InsightsConfig InsightsConfigAPI InsightsOnDemandDataGather InsightsRuntimeExtractor MachineAPIMigration MachineAPIOperatorDisableMachineHealthCheckController MachineAPIProviderOpenStack MachineConfigNodes MaxUnavailableStatefulSet MetricsCollectionProfiles MinimumKubeletVersion MixedCPUsAllocation MultiArchInstallAzure NodeSwap NutanixMultiSubnets OVNObservability OpenShiftPodSecurityAdmission PinnedImages PlatformOperators ProcMountType RouteAdvertisements RouteExternalCertificate ServiceAccountTokenNodeBinding SignatureStores SigstoreImageVerification StreamingCollectionEncodingToJSON StreamingCollectionEncodingToProtobuf TranslateStreamCloseWebsocketRequests UpgradeStatus UserNamespacesPodSecurityStandards UserNamespacesSupport VSphereMultiNetworks VolumeAttributesClass VolumeGroupSnapshot] 2025-12-04T11:45:10.796952198Z I1204 11:45:10.796145 1 event.go:377] Event(v1.ObjectReference{Kind:"Node", Namespace:"openshift-machine-config-operator", Name:"master-0", UID:"9618c86a-8a87-4c00-8dd2-9747684d64b3", APIVersion:"v1", ResourceVersion:"", FieldPath:""}): type: 'Normal' reason: 'FeatureGatesInitialized' FeatureGates updated to featuregates.Features{Enabled:[]v1.FeatureGateName{"AWSEFSDriverVolumeMetrics", "AdminNetworkPolicy", "AlibabaPlatform", "AzureWorkloadIdentity", "BareMetalLoadBalancer", "BuildCSIVolumes", "ChunkSizeMiB", "CloudDualStackNodeIPs", "DisableKubeletCloudCredentialProviders", "GCPLabelsTags", "HardwareSpeed", "IngressControllerLBSubnetsAWS", "KMSv1", "ManagedBootImages", "ManagedBootImagesAWS", "MultiArchInstallAWS", "MultiArchInstallGCP", "NetworkDiagnosticsConfig", "NetworkLiveMigration", "NetworkSegmentation", "NewOLM", "NodeDisruptionPolicy", "OnClusterBuild", "PersistentIPsForVirtualization", "PrivateHostedZoneAWS", "SetEIPForNLBIngressController", "VSphereControlPlaneMachineSet", "VSphereDriverConfiguration", "VSphereMultiVCenters", "VSphereStaticIPs", "ValidatingAdmissionPolicy"}, Disabled:[]v1.FeatureGateName{"AWSClusterHostedDNS", "AdditionalRoutingCapabilities", "AutomatedEtcdBackup", "BootcNodeManagement", "CSIDriverSharedResource", "ClusterAPIInstall", "ClusterAPIInstallIBMCloud", "ClusterMonitoringConfig", "ConsolePluginContentSecurityPolicy", "DNSNameResolver", "DynamicResourceAllocation", "EtcdBackendQuota", "EventedPLEG", "Example", "ExternalOIDC", "ExternalOIDCWithUIDAndExtraClaimMappings", "GCPClusterHostedDNS", "GatewayAPI", "ImageStreamImportMode", "IngressControllerDynamicConfigurationManager", "InsightsConfig", "InsightsConfigAPI", "InsightsOnDemandDataGather", "InsightsRuntimeExtractor", "MachineAPIMigration", "MachineAPIOperatorDisableMachineHealthCheckController", "MachineAPIProviderOpenStack", "MachineConfigNodes", "MaxUnavailableStatefulSet", "MetricsCollectionProfiles", "MinimumKubeletVersion", "MixedCPUsAllocation", "MultiArchInstallAzure", "NodeSwap", "NutanixMultiSubnets", "OVNObservability", "OpenShiftPodSecurityAdmission", "PinnedImages", "PlatformOperators", "ProcMountType", "RouteAdvertisements", "RouteExternalCertificate", "ServiceAccountTokenNodeBinding", "SignatureStores", "SigstoreImageVerification", "StreamingCollectionEncodingToJSON", "StreamingCollectionEncodingToProtobuf", "TranslateStreamCloseWebsocketRequests", "UpgradeStatus", "UserNamespacesPodSecurityStandards", "UserNamespacesSupport", "VSphereMultiNetworks", "VolumeAttributesClass", "VolumeGroupSnapshot"}} 2025-12-04T11:45:10.881872620Z I1204 11:45:10.881153 1 operator.go:389] On-cluster layering featuregate enabled, starting MachineOSConfig informer 2025-12-04T11:45:11.007885195Z I1204 11:45:11.003955 1 operator.go:436] Change observed to kube-apiserver-server-ca 2025-12-04T11:45:11.239863339Z I1204 11:45:11.239701 1 operator.go:436] Change observed to kube-apiserver-server-ca 2025-12-04T11:45:11.281454848Z I1204 11:45:11.281369 1 operator.go:416] Starting MachineConfigOperator 2025-12-04T11:45:11.281825839Z I1204 11:45:11.281788 1 sync.go:412] creating merged-trusted-image-registry-ca 2025-12-04T11:45:11.282405408Z I1204 11:45:11.282286 1 event.go:377] Event(v1.ObjectReference{Kind:"ClusterOperator", Namespace:"openshift-machine-config-operator", Name:"machine-config", UID:"cd635ec9-c3e7-4fed-aa22-37ef2b395ab1", APIVersion:"", ResourceVersion:"", FieldPath:""}): type: 'Normal' reason: 'OperatorVersionChanged' clusteroperator/machine-config started a version change from [] to [{operator 4.18.29} {operator-image quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b472823604757237c2d16bd6f6221f4cf562aa3b05942c7f602e1e8b2e55a7c6}] 2025-12-04T11:45:11.290907816Z E1204 11:45:11.288397 1 sync.go:2172] exiting image registry secrets fetch - machine-os-puller service account does not exist yet. 2025-12-04T11:45:11.290907816Z I1204 11:45:11.289176 1 sync.go:2256] MachineConfiguration object doesn't exist; a new one will be created 2025-12-04T11:45:24.828875970Z I1204 11:45:24.828734 1 sync.go:850] Skipping retry in ApplyManifests for error: Internal error occurred: admission plugin "authorization.openshift.io/RestrictSubjectBindings" failed to complete validation in 13s 2025-12-04T11:45:24.828875970Z I1204 11:45:24.828776 1 sync.go:984] Performing safety controllerconfig sync 2025-12-04T11:45:24.828875970Z E1204 11:45:24.828790 1 sync.go:192] Error performing safety controllerconfig sync: controllerconfig.machineconfiguration.openshift.io "machine-config-controller" not found 2025-12-04T11:45:24.829113678Z I1204 11:45:24.829052 1 event.go:377] Event(v1.ObjectReference{Kind:"ClusterOperator", Namespace:"openshift-machine-config-operator", Name:"machine-config", UID:"cd635ec9-c3e7-4fed-aa22-37ef2b395ab1", APIVersion:"", ResourceVersion:"", FieldPath:""}): type: 'Warning' reason: 'OperatorDegraded: MachineConfigDaemonFailed' Unable to apply 4.18.29: failed to apply machine config daemon manifests: Internal error occurred: admission plugin "authorization.openshift.io/RestrictSubjectBindings" failed to complete validation in 13s 2025-12-04T11:45:52.937696579Z E1204 11:45:52.937642 1 sync.go:2172] exiting image registry secrets fetch - machine-os-puller service account does not exist yet. 2025-12-04T11:45:52.937870695Z I1204 11:45:52.937687 1 event.go:377] Event(v1.ObjectReference{Kind:"ClusterOperator", Namespace:"openshift-machine-config-operator", Name:"machine-config", UID:"cd635ec9-c3e7-4fed-aa22-37ef2b395ab1", APIVersion:"", ResourceVersion:"", FieldPath:""}): type: 'Normal' reason: 'OperatorVersionChanged' clusteroperator/machine-config started a version change from [] to [{operator 4.18.29} {operator-image quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b472823604757237c2d16bd6f6221f4cf562aa3b05942c7f602e1e8b2e55a7c6}] 2025-12-04T11:45:52.938003530Z I1204 11:45:52.937987 1 sync.go:2288] Updating MachineConfiguration status 2025-12-04T11:51:45.703480187Z E1204 11:51:45.703383 1 leaderelection.go:429] Failed to update lock optimitically: Timeout: request did not complete within requested timeout - context deadline exceeded, falling back to slow path 2025-12-04T11:52:45.705734868Z E1204 11:52:45.705667 1 leaderelection.go:436] error retrieving resource lock openshift-machine-config-operator/machine-config: the server was unable to return a response in the time allotted, but may still be processing the request (get leases.coordination.k8s.io machine-config) 2025-12-04T11:52:58.701587519Z I1204 11:52:58.701501 1 leaderelection.go:297] failed to renew lease openshift-machine-config-operator/machine-config: timed out waiting for the condition 2025-12-04T11:53:32.704406828Z E1204 11:53:32.704312 1 leaderelection.go:322] Failed to release lock: Timeout: request did not complete within requested timeout - context deadline exceeded 2025-12-04T11:53:32.704406828Z I1204 11:53:32.704387 1 start.go:155] Stopped leading. Terminating. config/pod/openshift-machine-config-operator/machine-config-operator-dc5d7666f-p2cmn.json0000640000000000000000000001666015114274105027721 0ustar0000000000000000{"metadata":{"name":"machine-config-operator-dc5d7666f-p2cmn","generateName":"machine-config-operator-dc5d7666f-","namespace":"openshift-machine-config-operator","uid":"cdb50850-779d-4179-ba1d-65ffeb4bb2e9","resourceVersion":"11246","creationTimestamp":"2025-12-04T11:38:43Z","labels":{"k8s-app":"machine-config-operator","pod-template-hash":"dc5d7666f"},"annotations":{"k8s.ovn.org/pod-networks":"{\"default\":{\"ip_addresses\":[\"10.128.0.61/23\"],\"mac_address\":\"0a:58:0a:80:00:3d\",\"gateway_ips\":[\"10.128.0.1\"],\"routes\":[{\"dest\":\"10.128.0.0/16\",\"nextHop\":\"10.128.0.1\"},{\"dest\":\"172.30.0.0/16\",\"nextHop\":\"10.128.0.1\"},{\"dest\":\"169.254.0.5/32\",\"nextHop\":\"10.128.0.1\"},{\"dest\":\"100.64.0.0/16\",\"nextHop\":\"10.128.0.1\"}],\"ip_address\":\"10.128.0.61/23\",\"gateway_ip\":\"10.128.0.1\",\"role\":\"primary\"}}","k8s.v1.cni.cncf.io/network-status":"[{\n \"name\": \"ovn-kubernetes\",\n \"interface\": \"eth0\",\n \"ips\": [\n \"10.128.0.61\"\n ],\n \"mac\": \"0a:58:0a:80:00:3d\",\n \"default\": true,\n \"dns\": {}\n}]","openshift.io/required-scc":"anyuid","openshift.io/scc":"anyuid"},"ownerReferences":[{"apiVersion":"apps/v1","kind":"ReplicaSet","name":"machine-config-operator-dc5d7666f","uid":"218d0721-f46d-446e-a4a8-c022a975a753","controller":true,"blockOwnerDeletion":true}]},"spec":{"volumes":[{"name":"images","configMap":{"name":"machine-config-operator-images","defaultMode":420}},{"name":"proxy-tls","secret":{"secretName":"mco-proxy-tls","defaultMode":420}},{"name":"auth-proxy-config","configMap":{"name":"kube-rbac-proxy","defaultMode":420}},{"name":"kube-api-access-nhvdc","projected":{"sources":[{"serviceAccountToken":{"expirationSeconds":3607,"path":"token"}},{"configMap":{"name":"kube-root-ca.crt","items":[{"key":"ca.crt","path":"ca.crt"}]}},{"downwardAPI":{"items":[{"path":"namespace","fieldRef":{"apiVersion":"v1","fieldPath":"metadata.namespace"}}]}},{"configMap":{"name":"openshift-service-ca.crt","items":[{"key":"service-ca.crt","path":"service-ca.crt"}]}}],"defaultMode":420}}],"containers":[{"name":"machine-config-operator","image":"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b472823604757237c2d16bd6f6221f4cf562aa3b05942c7f602e1e8b2e55a7c6","args":["start","--images-json=/etc/mco/images/images.json","--payload-version=4.18.29","--operator-image=quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b472823604757237c2d16bd6f6221f4cf562aa3b05942c7f602e1e8b2e55a7c6"],"env":[{"name":"RELEASE_VERSION","value":"4.18.29"}],"resources":{"requests":{"cpu":"20m","memory":"50Mi"}},"volumeMounts":[{"name":"images","mountPath":"/etc/mco/images"},{"name":"kube-api-access-nhvdc","readOnly":true,"mountPath":"/var/run/secrets/kubernetes.io/serviceaccount"}],"terminationMessagePath":"/dev/termination-log","terminationMessagePolicy":"FallbackToLogsOnError","imagePullPolicy":"IfNotPresent","securityContext":{"capabilities":{"drop":["MKNOD"]}}},{"name":"kube-rbac-proxy","image":"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69ffd8f8dcceedc2d6eb306cea33f8beabc1be1308cd5f4ee8b9a8e3eab9843","args":["--secure-listen-address=0.0.0.0:9001","--config-file=/etc/kube-rbac-proxy/config-file.yaml","--tls-cipher-suites=TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305,TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305","--upstream=http://127.0.0.1:8797","--logtostderr=true","--tls-cert-file=/etc/tls/private/tls.crt","--tls-private-key-file=/etc/tls/private/tls.key"],"ports":[{"name":"metrics","containerPort":9001,"protocol":"TCP"}],"resources":{"requests":{"cpu":"20m","memory":"50Mi"}},"volumeMounts":[{"name":"proxy-tls","mountPath":"/etc/tls/private"},{"name":"auth-proxy-config","mountPath":"/etc/kube-rbac-proxy"},{"name":"kube-api-access-nhvdc","readOnly":true,"mountPath":"/var/run/secrets/kubernetes.io/serviceaccount"}],"terminationMessagePath":"/dev/termination-log","terminationMessagePolicy":"FallbackToLogsOnError","imagePullPolicy":"IfNotPresent","securityContext":{"capabilities":{"drop":["MKNOD"]}}}],"restartPolicy":"Always","terminationGracePeriodSeconds":30,"dnsPolicy":"ClusterFirst","nodeSelector":{"node-role.kubernetes.io/master":""},"serviceAccountName":"machine-config-operator","serviceAccount":"machine-config-operator","nodeName":"master-0","securityContext":{"seLinuxOptions":{"level":"s0:c18,c17"},"runAsUser":65534,"runAsNonRoot":true},"imagePullSecrets":[{"name":"machine-config-operator-dockercfg-l9v2g"}],"schedulerName":"default-scheduler","tolerations":[{"key":"node-role.kubernetes.io/master","operator":"Exists","effect":"NoSchedule"},{"key":"node.kubernetes.io/unreachable","operator":"Exists","effect":"NoExecute","tolerationSeconds":120},{"key":"node.kubernetes.io/not-ready","operator":"Exists","effect":"NoExecute","tolerationSeconds":120},{"key":"node.kubernetes.io/memory-pressure","operator":"Exists","effect":"NoSchedule"}],"priorityClassName":"system-cluster-critical","priority":2000000000,"enableServiceLinks":true,"preemptionPolicy":"PreemptLowerPriority"},"status":{"phase":"Running","conditions":[{"type":"PodReadyToStartContainers","status":"True","lastProbeTime":null,"lastTransitionTime":"2025-12-04T11:39:22Z"},{"type":"Initialized","status":"True","lastProbeTime":null,"lastTransitionTime":"2025-12-04T11:38:44Z"},{"type":"Ready","status":"True","lastProbeTime":null,"lastTransitionTime":"2025-12-04T11:53:34Z"},{"type":"ContainersReady","status":"True","lastProbeTime":null,"lastTransitionTime":"2025-12-04T11:53:34Z"},{"type":"PodScheduled","status":"True","lastProbeTime":null,"lastTransitionTime":"2025-12-04T11:38:43Z"}],"hostIP":"192.168.32.10","hostIPs":[{"ip":"192.168.32.10"}],"podIP":"10.128.0.61","podIPs":[{"ip":"10.128.0.61"}],"startTime":"2025-12-04T11:38:44Z","containerStatuses":[{"name":"kube-rbac-proxy","state":{"running":{"startedAt":"2025-12-04T11:39:21Z"}},"lastState":{},"ready":true,"restartCount":0,"image":"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69ffd8f8dcceedc2d6eb306cea33f8beabc1be1308cd5f4ee8b9a8e3eab9843","imageID":"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69ffd8f8dcceedc2d6eb306cea33f8beabc1be1308cd5f4ee8b9a8e3eab9843","containerID":"cri-o://21890527e3939ef8c71ed600b954fbd30bb4b48d4a2e73ffc24208dce81ed478","started":true,"volumeMounts":[{"name":"proxy-tls","mountPath":"/etc/tls/private"},{"name":"auth-proxy-config","mountPath":"/etc/kube-rbac-proxy"},{"name":"kube-api-access-nhvdc","mountPath":"/var/run/secrets/kubernetes.io/serviceaccount","readOnly":true,"recursiveReadOnly":"Disabled"}]},{"name":"machine-config-operator","state":{"running":{"startedAt":"2025-12-04T11:53:33Z"}},"lastState":{"terminated":{"exitCode":0,"reason":"Completed","startedAt":"2025-12-04T11:39:21Z","finishedAt":"2025-12-04T11:53:32Z","containerID":"cri-o://3bcbd231509913c4a508bf39ec4788d7227ea3b5ad2a02de0790fac547297cf4"}},"ready":true,"restartCount":1,"image":"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b472823604757237c2d16bd6f6221f4cf562aa3b05942c7f602e1e8b2e55a7c6","imageID":"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b472823604757237c2d16bd6f6221f4cf562aa3b05942c7f602e1e8b2e55a7c6","containerID":"cri-o://c5f5f1f7464567acecbfb64144a65775c557ceb3146468b7a5cd1ebd81a55aa4","started":true,"volumeMounts":[{"name":"images","mountPath":"/etc/mco/images"},{"name":"kube-api-access-nhvdc","mountPath":"/var/run/secrets/kubernetes.io/serviceaccount","readOnly":true,"recursiveReadOnly":"Disabled"}]}],"qosClass":"Burstable"}}kube-rbac-proxy-crio_current.log0000640000000000000000000000225215114274105034455 0ustar0000000000000000config/pod/openshift-machine-config-operator/logs/kube-rbac-proxy-crio-master-02025-12-04T11:36:06.937320150Z W1204 11:36:06.936977 1 deprecated.go:66] 2025-12-04T11:36:06.937320150Z ==== Removed Flag Warning ====================== 2025-12-04T11:36:06.937320150Z 2025-12-04T11:36:06.937320150Z logtostderr is removed in the k8s upstream and has no effect any more. 2025-12-04T11:36:06.937320150Z 2025-12-04T11:36:06.937320150Z =============================================== 2025-12-04T11:36:06.937320150Z 2025-12-04T11:36:06.937320150Z I1204 11:36:06.937208 1 kube-rbac-proxy.go:530] Reading config file: /etc/kubernetes/crio-metrics-proxy.cfg 2025-12-04T11:36:06.942236825Z I1204 11:36:06.942201 1 kube-rbac-proxy.go:233] Valid token audiences: 2025-12-04T11:36:06.942689238Z I1204 11:36:06.942646 1 dynamic_cafile_content.go:160] "Starting controller" name="client-ca::/etc/kubernetes/kubelet-ca.crt" 2025-12-04T11:36:06.944027687Z I1204 11:36:06.944007 1 kube-rbac-proxy.go:347] Reading certificate files 2025-12-04T11:36:06.944367917Z I1204 11:36:06.944306 1 kube-rbac-proxy.go:395] Starting TCP socket on :9637 2025-12-04T11:36:06.944901713Z I1204 11:36:06.944870 1 kube-rbac-proxy.go:402] Listening securely on :9637 config/pod/openshift-machine-config-operator/logs/kube-rbac-proxy-crio-master-0/setup_current.log0000640000000000000000000000012615114274105031726 0ustar00000000000000002025-12-04T11:35:17.669596430Z Waiting for kubelet key and certificate to be availablekube-rbac-proxy-crio_previous.log0000640000000000000000000000221015114274105034641 0ustar0000000000000000config/pod/openshift-machine-config-operator/logs/kube-rbac-proxy-crio-master-02025-12-04T11:35:39.874576169Z W1204 11:35:39.874235 1 deprecated.go:66] 2025-12-04T11:35:39.874576169Z ==== Removed Flag Warning ====================== 2025-12-04T11:35:39.874576169Z 2025-12-04T11:35:39.874576169Z logtostderr is removed in the k8s upstream and has no effect any more. 2025-12-04T11:35:39.874576169Z 2025-12-04T11:35:39.874576169Z =============================================== 2025-12-04T11:35:39.874576169Z 2025-12-04T11:35:39.874576169Z I1204 11:35:39.874399 1 kube-rbac-proxy.go:530] Reading config file: /etc/kubernetes/crio-metrics-proxy.cfg 2025-12-04T11:35:39.875573428Z E1204 11:35:39.875537 1 run.go:72] "command failed" err="failed to load kubeconfig: unable to build rest config based on provided path to kubeconfig file: invalid configuration: [unable to read client-cert /var/lib/kubelet/pki/kubelet-client-current.pem for default-auth due to open /var/lib/kubelet/pki/kubelet-client-current.pem: no such file or directory, unable to read client-key /var/lib/kubelet/pki/kubelet-client-current.pem for default-auth due to open /var/lib/kubelet/pki/kubelet-client-current.pem: no such file or directory]" config/pod/openshift-machine-config-operator/kube-rbac-proxy-crio-master-0.json0000640000000000000000000001513715114274105026260 0ustar0000000000000000{"metadata":{"name":"kube-rbac-proxy-crio-master-0","namespace":"openshift-machine-config-operator","uid":"95d981c0-feed-4340-9736-86859e667ed1","resourceVersion":"11926","creationTimestamp":"2025-12-04T11:37:31Z","annotations":{"kubernetes.io/config.hash":"3169f44496ed8a28c6d6a15511ab0eec","kubernetes.io/config.mirror":"3169f44496ed8a28c6d6a15511ab0eec","kubernetes.io/config.seen":"2025-12-04T11:35:14.483847236Z","kubernetes.io/config.source":"file","openshift.io/required-scc":"privileged","openshift.io/scc":"privileged","target.workload.openshift.io/management":"{\"effect\": \"PreferredDuringScheduling\"}"},"ownerReferences":[{"apiVersion":"v1","kind":"Node","name":"master-0","uid":"9618c86a-8a87-4c00-8dd2-9747684d64b3","controller":true}]},"spec":{"volumes":[{"name":"etc-kube","hostPath":{"path":"/etc/kubernetes","type":""}},{"name":"var-lib-kubelet","hostPath":{"path":"/var/lib/kubelet","type":""}}],"initContainers":[{"name":"setup","image":"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69ffd8f8dcceedc2d6eb306cea33f8beabc1be1308cd5f4ee8b9a8e3eab9843","command":["/bin/bash","-ec"],"args":["echo -n \"Waiting for kubelet key and certificate to be available\"\nwhile [ -n \"$(test -e /var/lib/kubelet/pki/kubelet-server-current.pem)\" ] ; do\n echo -n \".\"\n sleep 1\n (( tries += 1 ))\n if [[ \"${tries}\" -gt 10 ]]; then\n echo \"Timed out waiting for kubelet key and cert.\"\n exit 1\n fi\ndone\n"],"resources":{"requests":{"cpu":"5m","memory":"50Mi"}},"volumeMounts":[{"name":"var-lib-kubelet","mountPath":"/var","mountPropagation":"HostToContainer"}],"terminationMessagePath":"/dev/termination-log","terminationMessagePolicy":"FallbackToLogsOnError","imagePullPolicy":"IfNotPresent","securityContext":{"privileged":true}}],"containers":[{"name":"kube-rbac-proxy-crio","image":"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69ffd8f8dcceedc2d6eb306cea33f8beabc1be1308cd5f4ee8b9a8e3eab9843","args":["--secure-listen-address=:9637","--config-file=/etc/kubernetes/crio-metrics-proxy.cfg","--client-ca-file=/etc/kubernetes/kubelet-ca.crt","--logtostderr=true","--kubeconfig=/var/lib/kubelet/kubeconfig","--tls-cipher-suites=TLS_AES_128_GCM_SHA256,TLS_AES_256_GCM_SHA384,TLS_CHACHA20_POLY1305_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305_SHA256,TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305_SHA256","--tls-min-version=VersionTLS12","--upstream=http://127.0.0.1:9537","--tls-cert-file=/var/lib/kubelet/pki/kubelet-server-current.pem","--tls-private-key-file=/var/lib/kubelet/pki/kubelet-server-current.pem"],"ports":[{"hostPort":9637,"containerPort":9637,"protocol":"TCP"}],"resources":{"requests":{"cpu":"20m","memory":"50Mi"}},"volumeMounts":[{"name":"etc-kube","readOnly":true,"mountPath":"/etc/kubernetes","mountPropagation":"HostToContainer"},{"name":"var-lib-kubelet","readOnly":true,"mountPath":"/var/lib/kubelet","mountPropagation":"HostToContainer"}],"terminationMessagePath":"/dev/termination-log","terminationMessagePolicy":"FallbackToLogsOnError","imagePullPolicy":"IfNotPresent","securityContext":{"privileged":true}}],"restartPolicy":"Always","terminationGracePeriodSeconds":30,"dnsPolicy":"ClusterFirst","nodeName":"master-0","hostNetwork":true,"securityContext":{},"schedulerName":"default-scheduler","tolerations":[{"operator":"Exists","effect":"NoExecute"},{"key":"node.kubernetes.io/memory-pressure","operator":"Exists","effect":"NoSchedule"}],"priorityClassName":"system-cluster-critical","priority":2000000000,"enableServiceLinks":true,"preemptionPolicy":"PreemptLowerPriority"},"status":{"phase":"Running","conditions":[{"type":"PodReadyToStartContainers","status":"True","lastProbeTime":null,"lastTransitionTime":"2025-12-04T11:59:18Z"},{"type":"Initialized","status":"True","lastProbeTime":null,"lastTransitionTime":"2025-12-04T11:59:18Z"},{"type":"Ready","status":"True","lastProbeTime":null,"lastTransitionTime":"2025-12-04T11:59:18Z"},{"type":"ContainersReady","status":"True","lastProbeTime":null,"lastTransitionTime":"2025-12-04T11:59:18Z"},{"type":"PodScheduled","status":"True","lastProbeTime":null,"lastTransitionTime":"2025-12-04T11:59:18Z"}],"hostIP":"192.168.32.10","hostIPs":[{"ip":"192.168.32.10"}],"podIP":"192.168.32.10","podIPs":[{"ip":"192.168.32.10"}],"startTime":"2025-12-04T11:59:18Z","initContainerStatuses":[{"name":"setup","state":{"terminated":{"exitCode":0,"reason":"Completed","startedAt":"2025-12-04T11:35:17Z","finishedAt":"2025-12-04T11:35:17Z","containerID":"cri-o://fe978e3c9349858cb4986d59a0d297668af47052f5667b090a002d76ebcd619e"}},"lastState":{},"ready":true,"restartCount":0,"image":"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69ffd8f8dcceedc2d6eb306cea33f8beabc1be1308cd5f4ee8b9a8e3eab9843","imageID":"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69ffd8f8dcceedc2d6eb306cea33f8beabc1be1308cd5f4ee8b9a8e3eab9843","containerID":"cri-o://fe978e3c9349858cb4986d59a0d297668af47052f5667b090a002d76ebcd619e","started":false}],"containerStatuses":[{"name":"kube-rbac-proxy-crio","state":{"running":{"startedAt":"2025-12-04T11:36:06Z"}},"lastState":{"terminated":{"exitCode":1,"reason":"Error","message":"W1204 11:35:39.874235 1 deprecated.go:66] \n==== Removed Flag Warning ======================\n\nlogtostderr is removed in the k8s upstream and has no effect any more.\n\n===============================================\n\t\t\nI1204 11:35:39.874399 1 kube-rbac-proxy.go:530] Reading config file: /etc/kubernetes/crio-metrics-proxy.cfg\nE1204 11:35:39.875537 1 run.go:72] \"command failed\" err=\"failed to load kubeconfig: unable to build rest config based on provided path to kubeconfig file: invalid configuration: [unable to read client-cert /var/lib/kubelet/pki/kubelet-client-current.pem for default-auth due to open /var/lib/kubelet/pki/kubelet-client-current.pem: no such file or directory, unable to read client-key /var/lib/kubelet/pki/kubelet-client-current.pem for default-auth due to open /var/lib/kubelet/pki/kubelet-client-current.pem: no such file or directory]\"\n","startedAt":"2025-12-04T11:35:39Z","finishedAt":"2025-12-04T11:35:39Z","containerID":"cri-o://3ed798c104f2e301f0961443f8ab6df6e17ec4caae5f1436b2b93d415a28f796"}},"ready":true,"restartCount":3,"image":"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69ffd8f8dcceedc2d6eb306cea33f8beabc1be1308cd5f4ee8b9a8e3eab9843","imageID":"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69ffd8f8dcceedc2d6eb306cea33f8beabc1be1308cd5f4ee8b9a8e3eab9843","containerID":"cri-o://cdc388259d8085154ae8e01bcad3dcfe7d469aa6e6aee3f021b6d2209453f2f4","started":true}],"qosClass":"Burstable"}}kube-controller-manager-operator_current.log0000640000000000000000000001105015114274105044100 0ustar0000000000000000config/pod/openshift-kube-controller-manager-operator/logs/kube-controller-manager-operator-848f645654-7hmhg2025-12-04T11:55:45.448043929Z I1204 11:55:45.447799 1 cmd.go:253] Using service-serving-cert provided certificates 2025-12-04T11:55:45.448043929Z I1204 11:55:45.447971 1 leaderelection.go:121] The leader election gives 4 retries and allows for 30s of clock skew. The kube-apiserver downtime tolerance is 78s. Worst non-graceful lease acquisition is 2m43s. Worst graceful lease acquisition is {26s}. 2025-12-04T11:55:45.448483332Z I1204 11:55:45.448439 1 observer_polling.go:159] Starting file observer 2025-12-04T11:56:11.602299500Z I1204 11:56:11.602213 1 builder.go:304] kube-controller-manager-operator version 4.18.0-202511181540.p2.ga2cbb57.assembly.stream.el9-a2cbb57-a2cbb5725fe5e560ba40d884cfd079b3f8391f03 2025-12-04T11:56:26.271154526Z I1204 11:56:26.271094 1 secure_serving.go:57] Forcing use of http/1.1 only 2025-12-04T11:56:26.271227578Z W1204 11:56:26.271216 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected. 2025-12-04T11:56:26.271253789Z W1204 11:56:26.271244 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected. 2025-12-04T11:56:26.271279840Z W1204 11:56:26.271270 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected. 2025-12-04T11:56:26.271304721Z W1204 11:56:26.271295 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected. 2025-12-04T11:56:26.271327611Z W1204 11:56:26.271318 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected. 2025-12-04T11:56:26.271349972Z W1204 11:56:26.271341 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected. 2025-12-04T11:56:26.274758416Z I1204 11:56:26.274685 1 builder.go:446] detected SingleReplicaTopologyMode, the original leader election has been altered for the default SingleReplicaTopology 2025-12-04T11:56:26.275126127Z I1204 11:56:26.275035 1 leaderelection.go:254] attempting to acquire leader lease openshift-kube-controller-manager-operator/kube-controller-manager-operator-lock... 2025-12-04T11:56:26.282680236Z I1204 11:56:26.282622 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController 2025-12-04T11:56:26.282778779Z I1204 11:56:26.282745 1 secure_serving.go:213] Serving securely on [::]:8443 2025-12-04T11:56:26.282820171Z I1204 11:56:26.282759 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController 2025-12-04T11:56:26.282872062Z I1204 11:56:26.282849 1 tlsconfig.go:243] "Starting DynamicServingCertificateController" 2025-12-04T11:56:26.283006036Z I1204 11:56:26.282972 1 configmap_cafile_content.go:205] "Starting controller" name="client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file" 2025-12-04T11:56:26.283006036Z I1204 11:56:26.283002 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file 2025-12-04T11:56:26.283069968Z I1204 11:56:26.282828 1 dynamic_serving_content.go:135] "Starting controller" name="serving-cert::/var/run/secrets/serving-cert/tls.crt::/var/run/secrets/serving-cert/tls.key" 2025-12-04T11:56:26.283136250Z I1204 11:56:26.283078 1 configmap_cafile_content.go:205] "Starting controller" name="client-ca::kube-system::extension-apiserver-authentication::client-ca-file" 2025-12-04T11:56:26.283167841Z I1204 11:56:26.283147 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file 2025-12-04T11:56:26.385878190Z I1204 11:56:26.383878 1 shared_informer.go:320] Caches are synced for RequestHeaderAuthRequestController 2025-12-04T11:56:26.385878190Z I1204 11:56:26.383995 1 shared_informer.go:320] Caches are synced for client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file 2025-12-04T11:56:26.385878190Z I1204 11:56:26.384019 1 shared_informer.go:320] Caches are synced for client-ca::kube-system::extension-apiserver-authentication::client-ca-file 2025-12-04T11:59:28.940876096Z E1204 11:59:28.940760 1 leaderelection.go:436] error retrieving resource lock openshift-kube-controller-manager-operator/kube-controller-manager-operator-lock: Get "https://172.30.0.1:443/apis/coordination.k8s.io/v1/namespaces/openshift-kube-controller-manager-operator/leases/kube-controller-manager-operator-lock?timeout=4m0s": dial tcp 172.30.0.1:443: connect: connection refused kube-controller-manager-operator_previous.log0000640000000000000000000007500015114274105044277 0ustar0000000000000000config/pod/openshift-kube-controller-manager-operator/logs/kube-controller-manager-operator-848f645654-7hmhg2025-12-04T11:50:13.867839911Z I1204 11:50:13.867746 1 request.go:700] Waited for 1.580093717s due to client-side throttling, not priority and fairness, request: GET:https://172.30.0.1:443/api/v1/namespaces/openshift-kube-controller-manager/configmaps/config 2025-12-04T11:50:15.067913028Z I1204 11:50:15.067774 1 request.go:700] Waited for 1.39406098s due to client-side throttling, not priority and fairness, request: POST:https://172.30.0.1:443/api/v1/namespaces/openshift-kube-controller-manager/configmaps 2025-12-04T11:50:15.074846214Z I1204 11:50:15.074715 1 prune_controller.go:277] Nothing to prune 2025-12-04T11:50:15.075054500Z I1204 11:50:15.074987 1 event.go:377] Event(v1.ObjectReference{Kind:"Namespace", Namespace:"openshift-kube-controller-manager-operator", Name:"openshift-kube-controller-manager-operator", UID:"", APIVersion:"v1", ResourceVersion:"", FieldPath:""}): type: 'Normal' reason: 'ConfigMapCreated' Created ConfigMap/recycler-config-3 -n openshift-kube-controller-manager because it was missing 2025-12-04T11:50:15.470495110Z E1204 11:50:15.470438 1 base_controller.go:279] "Unhandled Error" err="SATokenSignerController reconciliation failed: unexpected addresses: 192.168.32.10" 2025-12-04T11:50:15.470915313Z I1204 11:50:15.470890 1 event.go:377] Event(v1.ObjectReference{Kind:"Namespace", Namespace:"openshift-kube-controller-manager-operator", Name:"openshift-kube-controller-manager-operator", UID:"", APIVersion:"v1", ResourceVersion:"", FieldPath:""}): type: 'Normal' reason: 'SATokenSignerControllerStuck' unexpected addresses: 192.168.32.10 2025-12-04T11:50:15.871798692Z I1204 11:50:15.871702 1 installer_controller.go:550] "master-0" is in transition to 2, but has not made progress because static pod is pending 2025-12-04T11:50:16.074345222Z I1204 11:50:16.074198 1 event.go:377] Event(v1.ObjectReference{Kind:"Namespace", Namespace:"openshift-kube-controller-manager-operator", Name:"openshift-kube-controller-manager-operator", UID:"", APIVersion:"v1", ResourceVersion:"", FieldPath:""}): type: 'Normal' reason: 'SecretCreated' Created Secret/service-account-private-key-3 -n openshift-kube-controller-manager because it was missing 2025-12-04T11:50:17.075086210Z I1204 11:50:17.075000 1 event.go:377] Event(v1.ObjectReference{Kind:"Namespace", Namespace:"openshift-kube-controller-manager-operator", Name:"openshift-kube-controller-manager-operator", UID:"", APIVersion:"v1", ResourceVersion:"", FieldPath:""}): type: 'Normal' reason: 'SecretCreated' Created Secret/serving-cert-3 -n openshift-kube-controller-manager because it was missing 2025-12-04T11:50:17.470275412Z I1204 11:50:17.470190 1 event.go:377] Event(v1.ObjectReference{Kind:"Namespace", Namespace:"openshift-kube-controller-manager-operator", Name:"openshift-kube-controller-manager-operator", UID:"", APIVersion:"v1", ResourceVersion:"", FieldPath:""}): type: 'Normal' reason: 'SATokenSignerControllerStuck' unexpected addresses: 192.168.32.10 2025-12-04T11:50:17.470368405Z E1204 11:50:17.470301 1 base_controller.go:279] "Unhandled Error" err="SATokenSignerController reconciliation failed: unexpected addresses: 192.168.32.10" 2025-12-04T11:50:18.074139894Z I1204 11:50:18.074049 1 event.go:377] Event(v1.ObjectReference{Kind:"Namespace", Namespace:"openshift-kube-controller-manager-operator", Name:"openshift-kube-controller-manager-operator", UID:"", APIVersion:"v1", ResourceVersion:"", FieldPath:""}): type: 'Normal' reason: 'SecretCreated' Created Secret/localhost-recovery-client-token-3 -n openshift-kube-controller-manager because it was missing 2025-12-04T11:50:18.870356380Z I1204 11:50:18.870263 1 installer_controller.go:550] "master-0" is in transition to 2, but has not made progress because static pod is pending 2025-12-04T11:50:19.075753548Z I1204 11:50:19.075634 1 prune_controller.go:277] Nothing to prune 2025-12-04T11:50:19.077683979Z I1204 11:50:19.077607 1 event.go:377] Event(v1.ObjectReference{Kind:"Namespace", Namespace:"openshift-kube-controller-manager-operator", Name:"openshift-kube-controller-manager-operator", UID:"", APIVersion:"v1", ResourceVersion:"", FieldPath:""}): type: 'Normal' reason: 'RevisionTriggered' new revision 3 triggered by "required configmap/kube-controller-manager-pod has changed,required configmap/serviceaccount-ca has changed" 2025-12-04T11:50:19.101331225Z I1204 11:50:19.100353 1 prune_controller.go:277] Nothing to prune 2025-12-04T11:50:19.471660213Z I1204 11:50:19.471566 1 event.go:377] Event(v1.ObjectReference{Kind:"Namespace", Namespace:"openshift-kube-controller-manager-operator", Name:"openshift-kube-controller-manager-operator", UID:"", APIVersion:"v1", ResourceVersion:"", FieldPath:""}): type: 'Normal' reason: 'SATokenSignerControllerStuck' unexpected addresses: 192.168.32.10 2025-12-04T11:50:19.471871189Z E1204 11:50:19.471833 1 base_controller.go:279] "Unhandled Error" err="SATokenSignerController reconciliation failed: unexpected addresses: 192.168.32.10" 2025-12-04T11:50:20.267276300Z I1204 11:50:20.267194 1 request.go:700] Waited for 1.166633446s due to client-side throttling, not priority and fairness, request: GET:https://172.30.0.1:443/api/v1/namespaces/openshift-kube-controller-manager/serviceaccounts/installer-sa 2025-12-04T11:50:21.482392175Z E1204 11:50:21.481190 1 base_controller.go:279] "Unhandled Error" err="SATokenSignerController reconciliation failed: unexpected addresses: 192.168.32.10" 2025-12-04T11:50:21.482392175Z I1204 11:50:21.481541 1 event.go:377] Event(v1.ObjectReference{Kind:"Namespace", Namespace:"openshift-kube-controller-manager-operator", Name:"openshift-kube-controller-manager-operator", UID:"", APIVersion:"v1", ResourceVersion:"", FieldPath:""}): type: 'Normal' reason: 'SATokenSignerControllerStuck' unexpected addresses: 192.168.32.10 2025-12-04T11:50:21.871964511Z I1204 11:50:21.871876 1 installer_controller.go:550] "master-0" is in transition to 2, but has not made progress because static pod is pending 2025-12-04T11:50:22.671121888Z I1204 11:50:22.671031 1 installer_controller.go:540] "master-0" moving to (v1.NodeStatus) { 2025-12-04T11:50:22.671121888Z NodeName: (string) (len=8) "master-0", 2025-12-04T11:50:22.671121888Z CurrentRevision: (int32) 0, 2025-12-04T11:50:22.671121888Z TargetRevision: (int32) 3, 2025-12-04T11:50:22.671121888Z LastFailedRevision: (int32) 0, 2025-12-04T11:50:22.671121888Z LastFailedTime: (*v1.Time)(), 2025-12-04T11:50:22.671121888Z LastFailedReason: (string) "", 2025-12-04T11:50:22.671121888Z LastFailedCount: (int) 0, 2025-12-04T11:50:22.671121888Z LastFallbackCount: (int) 0, 2025-12-04T11:50:22.671121888Z LastFailedRevisionErrors: ([]string) 2025-12-04T11:50:22.671121888Z } 2025-12-04T11:50:22.671121888Z because new revision pending 2025-12-04T11:50:22.672422439Z W1204 11:50:22.672368 1 dynamic_operator_client.go:352] .status.conditions["InstallerControllerDegraded"].reason is missing; this will eventually be fatal 2025-12-04T11:50:22.672422439Z W1204 11:50:22.672387 1 dynamic_operator_client.go:355] .status.conditions["InstallerControllerDegraded"].message is missing; this will eventually be fatal 2025-12-04T11:50:22.672422439Z W1204 11:50:22.672392 1 dynamic_operator_client.go:352] .status.conditions["NodeInstallerDegraded"].reason is missing; this will eventually be fatal 2025-12-04T11:50:22.672422439Z W1204 11:50:22.672397 1 dynamic_operator_client.go:355] .status.conditions["NodeInstallerDegraded"].message is missing; this will eventually be fatal 2025-12-04T11:50:22.672422439Z W1204 11:50:22.672400 1 dynamic_operator_client.go:352] .status.conditions["NodeInstallerProgressing"].reason is missing; this will eventually be fatal 2025-12-04T11:50:22.695874859Z I1204 11:50:22.695788 1 prune_controller.go:277] Nothing to prune 2025-12-04T11:50:22.698453040Z I1204 11:50:22.697492 1 status_controller.go:225] clusteroperator/kube-controller-manager diff {"status":{"conditions":[{"lastTransitionTime":"2025-12-04T11:37:45Z","message":"StaticPodsDegraded: pod/kube-controller-manager-master-0 container \"cluster-policy-controller\" is waiting: CrashLoopBackOff: back-off 1m20s restarting failed container=cluster-policy-controller pod=kube-controller-manager-master-0_openshift-kube-controller-manager(e6b437c60bb18680f4492b00b294e872)\nNodeControllerDegraded: All master nodes are ready","reason":"AsExpected","status":"False","type":"Degraded"},{"lastTransitionTime":"2025-12-04T11:38:24Z","message":"NodeInstallerProgressing: 1 node is at revision 0; 0 nodes have achieved new revision 3","reason":"NodeInstaller","status":"True","type":"Progressing"},{"lastTransitionTime":"2025-12-04T11:37:46Z","message":"StaticPodsAvailable: 0 nodes are active; 1 node is at revision 0; 0 nodes have achieved new revision 3","reason":"StaticPods_ZeroNodesActive","status":"False","type":"Available"},{"lastTransitionTime":"2025-12-04T11:37:45Z","message":"All is well","reason":"AsExpected","status":"True","type":"Upgradeable"},{"lastTransitionTime":"2025-12-04T11:37:44Z","reason":"NoData","status":"Unknown","type":"EvaluationConditionsDetected"}]}} 2025-12-04T11:50:22.711549178Z I1204 11:50:22.711375 1 event.go:377] Event(v1.ObjectReference{Kind:"Namespace", Namespace:"openshift-kube-controller-manager-operator", Name:"openshift-kube-controller-manager-operator", UID:"", APIVersion:"v1", ResourceVersion:"", FieldPath:""}): type: 'Normal' reason: 'OperatorStatusChanged' Status for clusteroperator/kube-controller-manager changed: Progressing message changed from "NodeInstallerProgressing: 1 node is at revision 0; 0 nodes have achieved new revision 2" to "NodeInstallerProgressing: 1 node is at revision 0; 0 nodes have achieved new revision 3",Available message changed from "StaticPodsAvailable: 0 nodes are active; 1 node is at revision 0; 0 nodes have achieved new revision 2" to "StaticPodsAvailable: 0 nodes are active; 1 node is at revision 0; 0 nodes have achieved new revision 3" 2025-12-04T11:50:23.071303066Z I1204 11:50:23.071169 1 event.go:377] Event(v1.ObjectReference{Kind:"Namespace", Namespace:"openshift-kube-controller-manager-operator", Name:"openshift-kube-controller-manager-operator", UID:"", APIVersion:"v1", ResourceVersion:"", FieldPath:""}): type: 'Normal' reason: 'SATokenSignerControllerStuck' unexpected addresses: 192.168.32.10 2025-12-04T11:50:23.071379418Z E1204 11:50:23.071299 1 base_controller.go:279] "Unhandled Error" err="SATokenSignerController reconciliation failed: unexpected addresses: 192.168.32.10" 2025-12-04T11:50:23.866972734Z I1204 11:50:23.866885 1 request.go:700] Waited for 1.162335722s due to client-side throttling, not priority and fairness, request: GET:https://172.30.0.1:443/api/v1/namespaces/openshift-kube-controller-manager/pods/installer-3-master-0 2025-12-04T11:50:24.679393254Z I1204 11:50:24.679092 1 event.go:377] Event(v1.ObjectReference{Kind:"Namespace", Namespace:"openshift-kube-controller-manager-operator", Name:"openshift-kube-controller-manager-operator", UID:"", APIVersion:"v1", ResourceVersion:"", FieldPath:""}): type: 'Normal' reason: 'PodCreated' Created Pod/installer-3-master-0 -n openshift-kube-controller-manager because it was missing 2025-12-04T11:50:25.072083467Z E1204 11:50:25.071676 1 base_controller.go:279] "Unhandled Error" err="SATokenSignerController reconciliation failed: unexpected addresses: 192.168.32.10" 2025-12-04T11:50:25.072083467Z I1204 11:50:25.071752 1 event.go:377] Event(v1.ObjectReference{Kind:"Namespace", Namespace:"openshift-kube-controller-manager-operator", Name:"openshift-kube-controller-manager-operator", UID:"", APIVersion:"v1", ResourceVersion:"", FieldPath:""}): type: 'Normal' reason: 'SATokenSignerControllerStuck' unexpected addresses: 192.168.32.10 2025-12-04T11:50:25.473703049Z I1204 11:50:25.473632 1 installer_controller.go:550] "master-0" is in transition to 3, but has not made progress because installer is not finished, but in Pending phase 2025-12-04T11:50:25.868027384Z I1204 11:50:25.867930 1 request.go:700] Waited for 1.188287199s due to client-side throttling, not priority and fairness, request: GET:https://172.30.0.1:443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-master-0 2025-12-04T11:50:27.067530031Z I1204 11:50:27.067436 1 request.go:700] Waited for 1.18931244s due to client-side throttling, not priority and fairness, request: GET:https://172.30.0.1:443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-master-0 2025-12-04T11:50:27.471831550Z I1204 11:50:27.471718 1 event.go:377] Event(v1.ObjectReference{Kind:"Namespace", Namespace:"openshift-kube-controller-manager-operator", Name:"openshift-kube-controller-manager-operator", UID:"", APIVersion:"v1", ResourceVersion:"", FieldPath:""}): type: 'Normal' reason: 'SATokenSignerControllerStuck' unexpected addresses: 192.168.32.10 2025-12-04T11:50:27.471913163Z E1204 11:50:27.471886 1 base_controller.go:279] "Unhandled Error" err="SATokenSignerController reconciliation failed: unexpected addresses: 192.168.32.10" 2025-12-04T11:50:27.871390197Z I1204 11:50:27.871303 1 installer_controller.go:550] "master-0" is in transition to 3, but has not made progress because installer is not finished, but in Running phase 2025-12-04T11:50:28.068003629Z I1204 11:50:28.067856 1 request.go:700] Waited for 1.19507819s due to client-side throttling, not priority and fairness, request: GET:https://172.30.0.1:443/api/v1/namespaces/openshift-kube-controller-manager/pods?labelSelector=app%3Dinstaller 2025-12-04T11:50:29.470215999Z I1204 11:50:29.470088 1 event.go:377] Event(v1.ObjectReference{Kind:"Namespace", Namespace:"openshift-kube-controller-manager-operator", Name:"openshift-kube-controller-manager-operator", UID:"", APIVersion:"v1", ResourceVersion:"", FieldPath:""}): type: 'Normal' reason: 'SATokenSignerControllerStuck' unexpected addresses: 192.168.32.10 2025-12-04T11:50:29.470414286Z E1204 11:50:29.470371 1 base_controller.go:279] "Unhandled Error" err="SATokenSignerController reconciliation failed: unexpected addresses: 192.168.32.10" 2025-12-04T11:50:29.870637832Z I1204 11:50:29.870560 1 installer_controller.go:550] "master-0" is in transition to 3, but has not made progress because installer is not finished, but in Running phase 2025-12-04T11:50:30.871599965Z E1204 11:50:30.871486 1 base_controller.go:279] "Unhandled Error" err="SATokenSignerController reconciliation failed: unexpected addresses: 192.168.32.10" 2025-12-04T11:50:30.872037659Z I1204 11:50:30.871341 1 event.go:377] Event(v1.ObjectReference{Kind:"Namespace", Namespace:"openshift-kube-controller-manager-operator", Name:"openshift-kube-controller-manager-operator", UID:"", APIVersion:"v1", ResourceVersion:"", FieldPath:""}): type: 'Normal' reason: 'SATokenSignerControllerStuck' unexpected addresses: 192.168.32.10 2025-12-04T11:50:32.069610514Z I1204 11:50:32.069487 1 event.go:377] Event(v1.ObjectReference{Kind:"Namespace", Namespace:"openshift-kube-controller-manager-operator", Name:"openshift-kube-controller-manager-operator", UID:"", APIVersion:"v1", ResourceVersion:"", FieldPath:""}): type: 'Normal' reason: 'SATokenSignerControllerStuck' unexpected addresses: 192.168.32.10 2025-12-04T11:50:32.069713417Z E1204 11:50:32.069671 1 base_controller.go:279] "Unhandled Error" err="SATokenSignerController reconciliation failed: unexpected addresses: 192.168.32.10" 2025-12-04T11:50:33.270985645Z I1204 11:50:33.270858 1 event.go:377] Event(v1.ObjectReference{Kind:"Namespace", Namespace:"openshift-kube-controller-manager-operator", Name:"openshift-kube-controller-manager-operator", UID:"", APIVersion:"v1", ResourceVersion:"", FieldPath:""}): type: 'Normal' reason: 'SATokenSignerControllerStuck' unexpected addresses: 192.168.32.10 2025-12-04T11:50:33.270985645Z E1204 11:50:33.270958 1 base_controller.go:279] "Unhandled Error" err="SATokenSignerController reconciliation failed: unexpected addresses: 192.168.32.10" 2025-12-04T11:50:39.481002566Z I1204 11:50:39.480870 1 event.go:377] Event(v1.ObjectReference{Kind:"Namespace", Namespace:"openshift-kube-controller-manager-operator", Name:"openshift-kube-controller-manager-operator", UID:"", APIVersion:"v1", ResourceVersion:"", FieldPath:""}): type: 'Normal' reason: 'SATokenSignerControllerStuck' unexpected addresses: 192.168.32.10 2025-12-04T11:50:39.482514851Z E1204 11:50:39.481112 1 base_controller.go:279] "Unhandled Error" err="SATokenSignerController reconciliation failed: unexpected addresses: 192.168.32.10" 2025-12-04T11:50:49.730605137Z I1204 11:50:49.730418 1 event.go:377] Event(v1.ObjectReference{Kind:"Namespace", Namespace:"openshift-kube-controller-manager-operator", Name:"openshift-kube-controller-manager-operator", UID:"", APIVersion:"v1", ResourceVersion:"", FieldPath:""}): type: 'Normal' reason: 'SATokenSignerControllerStuck' unexpected addresses: 192.168.32.10 2025-12-04T11:50:49.730605137Z E1204 11:50:49.730556 1 base_controller.go:279] "Unhandled Error" err="SATokenSignerController reconciliation failed: unexpected addresses: 192.168.32.10" 2025-12-04T11:51:44.463263205Z E1204 11:51:44.463162 1 leaderelection.go:429] Failed to update lock optimitically: Timeout: request did not complete within requested timeout - context deadline exceeded, falling back to slow path 2025-12-04T11:52:10.575337926Z E1204 11:52:10.575217 1 base_controller.go:279] "Unhandled Error" err="kube-controller-manager-InstallerState reconciliation failed: the server was unable to return a response in the time allotted, but may still be processing the request (get pods)" 2025-12-04T11:52:12.275749309Z E1204 11:52:12.275667 1 base_controller.go:279] "Unhandled Error" err="TargetConfigController reconciliation failed: the server was unable to return a response in the time allotted, but may still be processing the request" 2025-12-04T11:52:44.218338393Z E1204 11:52:44.218253 1 base_controller.go:279] "Unhandled Error" err="SATokenSignerController reconciliation failed: Timeout: request did not complete within requested timeout - context deadline exceeded" 2025-12-04T11:52:44.464603657Z E1204 11:52:44.464519 1 leaderelection.go:436] error retrieving resource lock openshift-kube-controller-manager-operator/kube-controller-manager-operator-lock: the server was unable to return a response in the time allotted, but may still be processing the request (get leases.coordination.k8s.io kube-controller-manager-operator-lock) 2025-12-04T11:52:44.582960146Z E1204 11:52:44.582796 1 base_controller.go:279] "Unhandled Error" err="kube-controller-manager-StaticPodState reconciliation failed: unable to ApplyStatus for operator using fieldManager \"kube-controller-manager-StaticPodState\": Timeout: request did not complete within requested timeout - context deadline exceeded" 2025-12-04T11:53:10.574585766Z I1204 11:53:10.574460 1 status_controller.go:225] clusteroperator/kube-controller-manager diff {"status":{"conditions":[{"lastTransitionTime":"2025-12-04T11:53:10Z","message":"StaticPodsDegraded: pod/kube-controller-manager-master-0 container \"cluster-policy-controller\" is waiting: CrashLoopBackOff: back-off 1m20s restarting failed container=cluster-policy-controller pod=kube-controller-manager-master-0_openshift-kube-controller-manager(e6b437c60bb18680f4492b00b294e872)","reason":"StaticPods_Error","status":"True","type":"Degraded"},{"lastTransitionTime":"2025-12-04T11:38:24Z","message":"NodeInstallerProgressing: 1 node is at revision 0; 0 nodes have achieved new revision 3","reason":"NodeInstaller","status":"True","type":"Progressing"},{"lastTransitionTime":"2025-12-04T11:37:46Z","message":"StaticPodsAvailable: 0 nodes are active; 1 node is at revision 0; 0 nodes have achieved new revision 3","reason":"StaticPods_ZeroNodesActive","status":"False","type":"Available"},{"lastTransitionTime":"2025-12-04T11:37:45Z","message":"All is well","reason":"AsExpected","status":"True","type":"Upgradeable"},{"lastTransitionTime":"2025-12-04T11:37:44Z","reason":"NoData","status":"Unknown","type":"EvaluationConditionsDetected"}]}} 2025-12-04T11:53:10.581113075Z E1204 11:53:10.581018 1 base_controller.go:279] "Unhandled Error" err="kube-controller-manager-InstallerState reconciliation failed: the server was unable to return a response in the time allotted, but may still be processing the request (get pods)" 2025-12-04T11:53:12.278521277Z E1204 11:53:12.278451 1 base_controller.go:279] "Unhandled Error" err="TargetConfigController reconciliation failed: the server was unable to return a response in the time allotted, but may still be processing the request" 2025-12-04T11:53:44.579969977Z E1204 11:53:44.579792 1 base_controller.go:279] "Unhandled Error" err="StatusSyncer_kube-controller-manager reconciliation failed: Timeout: request did not complete within requested timeout - context deadline exceeded" 2025-12-04T11:53:44.586702121Z E1204 11:53:44.586433 1 base_controller.go:279] "Unhandled Error" err="BackingResourceController-StaticResources reconciliation failed: [\"manifests/installer-sa.yaml\" (string): the server was unable to return a response in the time allotted, but may still be processing the request (get serviceaccounts installer-sa), \"manifests/installer-cluster-rolebinding.yaml\" (string): the server was unable to return a response in the time allotted, but may still be processing the request (get clusterrolebindings.rbac.authorization.k8s.io system:openshift:operator:openshift-kube-controller-manager-installer), unable to ApplyStatus for operator using fieldManager \"BackingResourceController-StaticResources\": Timeout: request did not complete within requested timeout - context deadline exceeded]" 2025-12-04T11:53:44.587737555Z I1204 11:53:44.587671 1 status_controller.go:225] clusteroperator/kube-controller-manager diff {"status":{"conditions":[{"lastTransitionTime":"2025-12-04T11:53:44Z","message":"StaticPodsDegraded: pod/kube-controller-manager-master-0 container \"cluster-policy-controller\" is waiting: CrashLoopBackOff: back-off 1m20s restarting failed container=cluster-policy-controller pod=kube-controller-manager-master-0_openshift-kube-controller-manager(e6b437c60bb18680f4492b00b294e872)","reason":"StaticPods_Error","status":"True","type":"Degraded"},{"lastTransitionTime":"2025-12-04T11:38:24Z","message":"NodeInstallerProgressing: 1 node is at revision 0; 0 nodes have achieved new revision 3","reason":"NodeInstaller","status":"True","type":"Progressing"},{"lastTransitionTime":"2025-12-04T11:37:46Z","message":"StaticPodsAvailable: 0 nodes are active; 1 node is at revision 0; 0 nodes have achieved new revision 3","reason":"StaticPods_ZeroNodesActive","status":"False","type":"Available"},{"lastTransitionTime":"2025-12-04T11:37:45Z","message":"All is well","reason":"AsExpected","status":"True","type":"Upgradeable"},{"lastTransitionTime":"2025-12-04T11:37:44Z","reason":"NoData","status":"Unknown","type":"EvaluationConditionsDetected"}]}} 2025-12-04T11:54:10.584381123Z E1204 11:54:10.584258 1 base_controller.go:279] "Unhandled Error" err="kube-controller-manager-InstallerState reconciliation failed: the server was unable to return a response in the time allotted, but may still be processing the request (get pods)" 2025-12-04T11:54:12.281993252Z E1204 11:54:12.281795 1 base_controller.go:279] "Unhandled Error" err="TargetConfigController reconciliation failed: the server was unable to return a response in the time allotted, but may still be processing the request" 2025-12-04T11:54:18.226401277Z E1204 11:54:18.226305 1 base_controller.go:279] "Unhandled Error" err="SATokenSignerController reconciliation failed: Timeout: request did not complete within requested timeout - context deadline exceeded" 2025-12-04T11:54:18.469390886Z E1204 11:54:18.469311 1 leaderelection.go:429] Failed to update lock optimitically: Timeout: request did not complete within requested timeout - context deadline exceeded, falling back to slow path 2025-12-04T11:54:18.591108953Z E1204 11:54:18.591044 1 base_controller.go:279] "Unhandled Error" err="kube-controller-manager-StaticPodState reconciliation failed: unable to ApplyStatus for operator using fieldManager \"kube-controller-manager-StaticPodState\": Timeout: request did not complete within requested timeout - context deadline exceeded" 2025-12-04T11:54:18.591366541Z E1204 11:54:18.591340 1 base_controller.go:279] "Unhandled Error" err="StatusSyncer_kube-controller-manager reconciliation failed: Timeout: request did not complete within requested timeout - context deadline exceeded" 2025-12-04T11:54:18.593225030Z I1204 11:54:18.593195 1 status_controller.go:225] clusteroperator/kube-controller-manager diff {"status":{"conditions":[{"lastTransitionTime":"2025-12-04T11:54:18Z","message":"StaticPodsDegraded: pod/kube-controller-manager-master-0 container \"cluster-policy-controller\" is waiting: CrashLoopBackOff: back-off 1m20s restarting failed container=cluster-policy-controller pod=kube-controller-manager-master-0_openshift-kube-controller-manager(e6b437c60bb18680f4492b00b294e872)","reason":"StaticPods_Error","status":"True","type":"Degraded"},{"lastTransitionTime":"2025-12-04T11:38:24Z","message":"NodeInstallerProgressing: 1 node is at revision 0; 0 nodes have achieved new revision 3","reason":"NodeInstaller","status":"True","type":"Progressing"},{"lastTransitionTime":"2025-12-04T11:37:46Z","message":"StaticPodsAvailable: 0 nodes are active; 1 node is at revision 0; 0 nodes have achieved new revision 3","reason":"StaticPods_ZeroNodesActive","status":"False","type":"Available"},{"lastTransitionTime":"2025-12-04T11:37:45Z","message":"All is well","reason":"AsExpected","status":"True","type":"Upgradeable"},{"lastTransitionTime":"2025-12-04T11:37:44Z","reason":"NoData","status":"Unknown","type":"EvaluationConditionsDetected"}]}} 2025-12-04T11:54:52.597036513Z E1204 11:54:52.596942 1 base_controller.go:279] "Unhandled Error" err="StatusSyncer_kube-controller-manager reconciliation failed: Timeout: request did not complete within requested timeout - context deadline exceeded" 2025-12-04T11:54:52.598983702Z I1204 11:54:52.598912 1 status_controller.go:225] clusteroperator/kube-controller-manager diff {"status":{"conditions":[{"lastTransitionTime":"2025-12-04T11:54:52Z","message":"StaticPodsDegraded: pod/kube-controller-manager-master-0 container \"cluster-policy-controller\" is waiting: CrashLoopBackOff: back-off 1m20s restarting failed container=cluster-policy-controller pod=kube-controller-manager-master-0_openshift-kube-controller-manager(e6b437c60bb18680f4492b00b294e872)","reason":"StaticPods_Error","status":"True","type":"Degraded"},{"lastTransitionTime":"2025-12-04T11:38:24Z","message":"NodeInstallerProgressing: 1 node is at revision 0; 0 nodes have achieved new revision 3","reason":"NodeInstaller","status":"True","type":"Progressing"},{"lastTransitionTime":"2025-12-04T11:37:46Z","message":"StaticPodsAvailable: 0 nodes are active; 1 node is at revision 0; 0 nodes have achieved new revision 3","reason":"StaticPods_ZeroNodesActive","status":"False","type":"Available"},{"lastTransitionTime":"2025-12-04T11:37:45Z","message":"All is well","reason":"AsExpected","status":"True","type":"Upgradeable"},{"lastTransitionTime":"2025-12-04T11:37:44Z","reason":"NoData","status":"Unknown","type":"EvaluationConditionsDetected"}]}} 2025-12-04T11:55:10.460598222Z E1204 11:55:10.459989 1 leaderelection.go:436] error retrieving resource lock openshift-kube-controller-manager-operator/kube-controller-manager-operator-lock: Get "https://172.30.0.1:443/apis/coordination.k8s.io/v1/namespaces/openshift-kube-controller-manager-operator/leases/kube-controller-manager-operator-lock?timeout=4m0s": context deadline exceeded 2025-12-04T11:55:10.460675594Z I1204 11:55:10.460641 1 leaderelection.go:297] failed to renew lease openshift-kube-controller-manager-operator/kube-controller-manager-operator-lock: timed out waiting for the condition 2025-12-04T11:55:10.587562888Z E1204 11:55:10.587485 1 base_controller.go:279] "Unhandled Error" err="kube-controller-manager-InstallerState reconciliation failed: the server was unable to return a response in the time allotted, but may still be processing the request (get pods)" 2025-12-04T11:55:12.285153811Z E1204 11:55:12.285054 1 base_controller.go:279] "Unhandled Error" err="TargetConfigController reconciliation failed: the server was unable to return a response in the time allotted, but may still be processing the request" 2025-12-04T11:55:26.606885933Z E1204 11:55:26.604237 1 base_controller.go:279] "Unhandled Error" err="StatusSyncer_kube-controller-manager reconciliation failed: Timeout: request did not complete within requested timeout - context deadline exceeded" 2025-12-04T11:55:26.610859703Z I1204 11:55:26.608671 1 status_controller.go:225] clusteroperator/kube-controller-manager diff {"status":{"conditions":[{"lastTransitionTime":"2025-12-04T11:55:26Z","message":"StaticPodsDegraded: pod/kube-controller-manager-master-0 container \"cluster-policy-controller\" is waiting: CrashLoopBackOff: back-off 1m20s restarting failed container=cluster-policy-controller pod=kube-controller-manager-master-0_openshift-kube-controller-manager(e6b437c60bb18680f4492b00b294e872)","reason":"StaticPods_Error","status":"True","type":"Degraded"},{"lastTransitionTime":"2025-12-04T11:38:24Z","message":"NodeInstallerProgressing: 1 node is at revision 0; 0 nodes have achieved new revision 3","reason":"NodeInstaller","status":"True","type":"Progressing"},{"lastTransitionTime":"2025-12-04T11:37:46Z","message":"StaticPodsAvailable: 0 nodes are active; 1 node is at revision 0; 0 nodes have achieved new revision 3","reason":"StaticPods_ZeroNodesActive","status":"False","type":"Available"},{"lastTransitionTime":"2025-12-04T11:37:45Z","message":"All is well","reason":"AsExpected","status":"True","type":"Upgradeable"},{"lastTransitionTime":"2025-12-04T11:37:44Z","reason":"NoData","status":"Unknown","type":"EvaluationConditionsDetected"}]}} 2025-12-04T11:55:30.133852567Z I1204 11:55:30.133759 1 prune_controller.go:277] Nothing to prune 2025-12-04T11:55:30.133949120Z E1204 11:55:30.133906 1 base_controller.go:279] "Unhandled Error" err="SATokenSignerController reconciliation failed: the server was unable to return a response in the time allotted, but may still be processing the request (get pods)" 2025-12-04T11:55:30.598713774Z I1204 11:55:30.598558 1 prune_controller.go:277] Nothing to prune 2025-12-04T11:55:44.464609034Z E1204 11:55:44.464507 1 leaderelection.go:322] Failed to release lock: Timeout: request did not complete within requested timeout - context deadline exceeded 2025-12-04T11:55:44.464697096Z W1204 11:55:44.464604 1 leaderelection.go:84] leader election lost kube-controller-manager-operator-848f645654-7hmhg.json0000640000000000000000000001451115114274105033217 0ustar0000000000000000config/pod/openshift-kube-controller-manager-operator{"metadata":{"name":"kube-controller-manager-operator-848f645654-7hmhg","generateName":"kube-controller-manager-operator-848f645654-","namespace":"openshift-kube-controller-manager-operator","uid":"bac00ac5-0371-48fd-b1cc-60438f08ae94","resourceVersion":"11273","creationTimestamp":"2025-12-04T11:35:49Z","labels":{"app":"kube-controller-manager-operator","pod-template-hash":"848f645654"},"annotations":{"k8s.ovn.org/pod-networks":"{\"default\":{\"ip_addresses\":[\"10.128.0.16/23\"],\"mac_address\":\"0a:58:0a:80:00:10\",\"gateway_ips\":[\"10.128.0.1\"],\"routes\":[{\"dest\":\"10.128.0.0/16\",\"nextHop\":\"10.128.0.1\"},{\"dest\":\"172.30.0.0/16\",\"nextHop\":\"10.128.0.1\"},{\"dest\":\"169.254.0.5/32\",\"nextHop\":\"10.128.0.1\"},{\"dest\":\"100.64.0.0/16\",\"nextHop\":\"10.128.0.1\"}],\"ip_address\":\"10.128.0.16/23\",\"gateway_ip\":\"10.128.0.1\",\"role\":\"primary\"}}","k8s.v1.cni.cncf.io/network-status":"[{\n \"name\": \"ovn-kubernetes\",\n \"interface\": \"eth0\",\n \"ips\": [\n \"10.128.0.16\"\n ],\n \"mac\": \"0a:58:0a:80:00:10\",\n \"default\": true,\n \"dns\": {}\n}]"},"ownerReferences":[{"apiVersion":"apps/v1","kind":"ReplicaSet","name":"kube-controller-manager-operator-848f645654","uid":"fdb82690-13b9-4d11-b067-83be72428efa","controller":true,"blockOwnerDeletion":true}]},"spec":{"volumes":[{"name":"serving-cert","secret":{"secretName":"kube-controller-manager-operator-serving-cert","defaultMode":420,"optional":true}},{"name":"config","configMap":{"name":"kube-controller-manager-operator-config","defaultMode":420}},{"name":"kube-api-access","projected":{"sources":[{"serviceAccountToken":{"expirationSeconds":3600,"path":"token"}},{"configMap":{"name":"kube-root-ca.crt","items":[{"key":"ca.crt","path":"ca.crt"}]}},{"downwardAPI":{"items":[{"path":"namespace","fieldRef":{"apiVersion":"v1","fieldPath":"metadata.namespace"}}]}}],"defaultMode":420}}],"containers":[{"name":"kube-controller-manager-operator","image":"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a824e468cf8dd61d347e35b2ee5bc2f815666957647098e21a1bb56ff613e5b9","command":["cluster-kube-controller-manager-operator","operator"],"args":["--config=/var/run/configmaps/config/config.yaml"],"ports":[{"name":"metrics","containerPort":8443,"protocol":"TCP"}],"env":[{"name":"IMAGE","value":"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:6d5891cdd7dcf7c9081de8b364b4c96446b7f946f7880fbae291a4592a198264"},{"name":"OPERATOR_IMAGE","value":"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a824e468cf8dd61d347e35b2ee5bc2f815666957647098e21a1bb56ff613e5b9"},{"name":"CLUSTER_POLICY_CONTROLLER_IMAGE","value":"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d64c13fe7663a0b4ae61d103b1b7598adcf317a01826f296bcb66b1a2de83c96"},{"name":"TOOLS_IMAGE","value":"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fa40d32981d88f32a9cdedc4cdd4a08c43e0d17bd4cc3fc3a87e9d1c7e1259d0"},{"name":"OPERATOR_IMAGE_VERSION","value":"4.18.29"},{"name":"OPERAND_IMAGE_VERSION","value":"1.31.13"},{"name":"POD_NAME","valueFrom":{"fieldRef":{"apiVersion":"v1","fieldPath":"metadata.name"}}}],"resources":{"requests":{"cpu":"10m","memory":"50Mi"}},"volumeMounts":[{"name":"config","mountPath":"/var/run/configmaps/config"},{"name":"serving-cert","mountPath":"/var/run/secrets/serving-cert"},{"name":"kube-api-access","readOnly":true,"mountPath":"/var/run/secrets/kubernetes.io/serviceaccount"}],"terminationMessagePath":"/dev/termination-log","terminationMessagePolicy":"FallbackToLogsOnError","imagePullPolicy":"IfNotPresent","securityContext":{"capabilities":{"drop":["ALL"]},"allowPrivilegeEscalation":false}}],"restartPolicy":"Always","terminationGracePeriodSeconds":30,"dnsPolicy":"ClusterFirst","nodeSelector":{"node-role.kubernetes.io/master":""},"serviceAccountName":"kube-controller-manager-operator","serviceAccount":"kube-controller-manager-operator","automountServiceAccountToken":false,"nodeName":"master-0","securityContext":{"runAsUser":65534,"runAsNonRoot":true,"seccompProfile":{"type":"RuntimeDefault"}},"schedulerName":"default-scheduler","tolerations":[{"key":"node-role.kubernetes.io/master","operator":"Exists","effect":"NoSchedule"},{"key":"node-role.kubernetes.io/control-plane","operator":"Exists","effect":"NoExecute"},{"key":"node.kubernetes.io/unreachable","operator":"Exists","effect":"NoExecute","tolerationSeconds":120},{"key":"node.kubernetes.io/not-ready","operator":"Exists","effect":"NoExecute","tolerationSeconds":120},{"key":"node.kubernetes.io/memory-pressure","operator":"Exists","effect":"NoSchedule"}],"priorityClassName":"system-cluster-critical","priority":2000000000,"enableServiceLinks":true,"preemptionPolicy":"PreemptLowerPriority"},"status":{"phase":"Running","conditions":[{"type":"PodReadyToStartContainers","status":"True","lastProbeTime":null,"lastTransitionTime":"2025-12-04T11:37:43Z"},{"type":"Initialized","status":"True","lastProbeTime":null,"lastTransitionTime":"2025-12-04T11:37:30Z"},{"type":"Ready","status":"True","lastProbeTime":null,"lastTransitionTime":"2025-12-04T11:55:46Z"},{"type":"ContainersReady","status":"True","lastProbeTime":null,"lastTransitionTime":"2025-12-04T11:55:46Z"},{"type":"PodScheduled","status":"True","lastProbeTime":null,"lastTransitionTime":"2025-12-04T11:37:30Z"}],"hostIP":"192.168.32.10","hostIPs":[{"ip":"192.168.32.10"}],"podIP":"10.128.0.16","podIPs":[{"ip":"10.128.0.16"}],"startTime":"2025-12-04T11:37:30Z","containerStatuses":[{"name":"kube-controller-manager-operator","state":{"running":{"startedAt":"2025-12-04T11:55:45Z"}},"lastState":{"terminated":{"exitCode":0,"reason":"Completed","startedAt":"2025-12-04T11:43:20Z","finishedAt":"2025-12-04T11:55:44Z","containerID":"cri-o://e0fdeac9876b7627a8946145261c72d16f488a3fb3a2e313379790f4879d5017"}},"ready":true,"restartCount":3,"image":"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a824e468cf8dd61d347e35b2ee5bc2f815666957647098e21a1bb56ff613e5b9","imageID":"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a824e468cf8dd61d347e35b2ee5bc2f815666957647098e21a1bb56ff613e5b9","containerID":"cri-o://ecd7f604588d7f724b04cf8225b65a6a6b0f7f44acb72826e4d9692d423df77a","started":true,"volumeMounts":[{"name":"config","mountPath":"/var/run/configmaps/config"},{"name":"serving-cert","mountPath":"/var/run/secrets/serving-cert"},{"name":"kube-api-access","mountPath":"/var/run/secrets/kubernetes.io/serviceaccount","readOnly":true,"recursiveReadOnly":"Disabled"}]}],"qosClass":"Burstable"}}kube-controller-manager-recovery-controller_current.log0000640000000000000000000011243015114274105042037 0ustar0000000000000000config/pod/openshift-kube-controller-manager/logs/kube-controller-manager-master-02025-12-04T11:45:45.790171313Z + timeout 3m /bin/bash -exuo pipefail -c 'while [ -n "$(ss -Htanop \( sport = 9443 \))" ]; do sleep 1; done' 2025-12-04T11:45:45.796011523Z ++ ss -Htanop '(' sport = 9443 ')' 2025-12-04T11:45:45.809383140Z + '[' -n '' ']' 2025-12-04T11:45:45.810383373Z + exec cluster-kube-controller-manager-operator cert-recovery-controller --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/kube-controller-cert-syncer-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager --listen=0.0.0.0:9443 -v=2 2025-12-04T11:45:45.859860918Z W1204 11:45:45.859583 1 cmd.go:257] Using insecure, self-signed certificates 2025-12-04T11:45:45.860120297Z I1204 11:45:45.860084 1 crypto.go:601] Generating new CA for cert-recovery-controller-signer@1764848745 cert, and key in /tmp/serving-cert-4108366648/serving-signer.crt, /tmp/serving-cert-4108366648/serving-signer.key 2025-12-04T11:45:46.425130074Z I1204 11:45:46.424977 1 leaderelection.go:121] The leader election gives 4 retries and allows for 30s of clock skew. The kube-apiserver downtime tolerance is 78s. Worst non-graceful lease acquisition is 2m43s. Worst graceful lease acquisition is {26s}. 2025-12-04T11:45:46.426026574Z I1204 11:45:46.425939 1 observer_polling.go:159] Starting file observer 2025-12-04T11:45:46.426178859Z I1204 11:45:46.426142 1 envvar.go:172] "Feature gate default state" feature="InformerResourceVersion" enabled=false 2025-12-04T11:45:46.426178859Z I1204 11:45:46.426161 1 envvar.go:172] "Feature gate default state" feature="WatchListClient" enabled=false 2025-12-04T11:45:46.430848031Z W1204 11:45:46.430717 1 builder.go:272] unable to get owner reference (falling back to namespace): Get "https://localhost:6443/api/v1/namespaces/openshift-kube-controller-manager/pods": tls: failed to verify certificate: x509: certificate is valid for kubernetes, kubernetes.default, kubernetes.default.svc, kubernetes.default.svc.cluster.local, openshift, openshift.default, openshift.default.svc, openshift.default.svc.cluster.local, not localhost-recovery 2025-12-04T11:45:46.431063558Z I1204 11:45:46.431023 1 builder.go:304] cert-recovery-controller version 4.18.0-202511181540.p2.ga2cbb57.assembly.stream.el9-a2cbb57-a2cbb5725fe5e560ba40d884cfd079b3f8391f03 2025-12-04T11:45:46.434601324Z W1204 11:45:46.434546 1 builder.go:364] unable to get control plane topology, using HA cluster values for leader election: Get "https://localhost:6443/apis/config.openshift.io/v1/infrastructures/cluster": tls: failed to verify certificate: x509: certificate is valid for kubernetes, kubernetes.default, kubernetes.default.svc, kubernetes.default.svc.cluster.local, openshift, openshift.default, openshift.default.svc, openshift.default.svc.cluster.local, not localhost-recovery 2025-12-04T11:45:46.434795190Z I1204 11:45:46.434695 1 event.go:377] Event(v1.ObjectReference{Kind:"Namespace", Namespace:"openshift-kube-controller-manager", Name:"openshift-kube-controller-manager", UID:"", APIVersion:"v1", ResourceVersion:"", FieldPath:""}): type: 'Warning' reason: 'ControlPlaneTopology' unable to get control plane topology, using HA cluster values for leader election: Get "https://localhost:6443/apis/config.openshift.io/v1/infrastructures/cluster": tls: failed to verify certificate: x509: certificate is valid for kubernetes, kubernetes.default, kubernetes.default.svc, kubernetes.default.svc.cluster.local, openshift, openshift.default, openshift.default.svc, openshift.default.svc.cluster.local, not localhost-recovery 2025-12-04T11:45:46.434937575Z I1204 11:45:46.434903 1 leaderelection.go:254] attempting to acquire leader lease openshift-kube-controller-manager/cert-recovery-controller-lock... 2025-12-04T11:45:46.437993295Z E1204 11:45:46.437770 1 leaderelection.go:436] error retrieving resource lock openshift-kube-controller-manager/cert-recovery-controller-lock: Get "https://localhost:6443/apis/coordination.k8s.io/v1/namespaces/openshift-kube-controller-manager/leases/cert-recovery-controller-lock?timeout=1m47s": tls: failed to verify certificate: x509: certificate is valid for kubernetes, kubernetes.default, kubernetes.default.svc, kubernetes.default.svc.cluster.local, openshift, openshift.default, openshift.default.svc, openshift.default.svc.cluster.local, not localhost-recovery 2025-12-04T11:45:46.439536594Z E1204 11:45:46.439473 1 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://localhost:6443/api/v1/namespaces/openshift-kube-controller-manager/events\": tls: failed to verify certificate: x509: certificate is valid for kubernetes, kubernetes.default, kubernetes.default.svc, kubernetes.default.svc.cluster.local, openshift, openshift.default, openshift.default.svc, openshift.default.svc.cluster.local, not localhost-recovery" event="&Event{ObjectMeta:{openshift-kube-controller-manager.187e00959585f6b3 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Namespace,Namespace:openshift-kube-controller-manager,Name:openshift-kube-controller-manager,UID:,APIVersion:v1,ResourceVersion:,FieldPath:,},Reason:ControlPlaneTopology,Message:unable to get control plane topology, using HA cluster values for leader election: Get \"https://localhost:6443/apis/config.openshift.io/v1/infrastructures/cluster\": tls: failed to verify certificate: x509: certificate is valid for kubernetes, kubernetes.default, kubernetes.default.svc, kubernetes.default.svc.cluster.local, openshift, openshift.default, openshift.default.svc, openshift.default.svc.cluster.local, not localhost-recovery,Source:EventSource{Component:cert-recovery-controller,Host:,},FirstTimestamp:2025-12-04 11:45:46.434524851 +0000 UTC m=+0.619194388,LastTimestamp:2025-12-04 11:45:46.434524851 +0000 UTC m=+0.619194388,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:cert-recovery-controller,ReportingInstance:,}" 2025-12-04T11:45:47.904385742Z E1204 11:45:47.904275 1 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://localhost:6443/api/v1/namespaces/openshift-kube-controller-manager/events\": tls: failed to verify certificate: x509: certificate is valid for kubernetes, kubernetes.default, kubernetes.default.svc, kubernetes.default.svc.cluster.local, openshift, openshift.default, openshift.default.svc, openshift.default.svc.cluster.local, not localhost-recovery" event="&Event{ObjectMeta:{openshift-kube-controller-manager.187e00959585f6b3 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Namespace,Namespace:openshift-kube-controller-manager,Name:openshift-kube-controller-manager,UID:,APIVersion:v1,ResourceVersion:,FieldPath:,},Reason:ControlPlaneTopology,Message:unable to get control plane topology, using HA cluster values for leader election: Get \"https://localhost:6443/apis/config.openshift.io/v1/infrastructures/cluster\": tls: failed to verify certificate: x509: certificate is valid for kubernetes, kubernetes.default, kubernetes.default.svc, kubernetes.default.svc.cluster.local, openshift, openshift.default, openshift.default.svc, openshift.default.svc.cluster.local, not localhost-recovery,Source:EventSource{Component:cert-recovery-controller,Host:,},FirstTimestamp:2025-12-04 11:45:46.434524851 +0000 UTC m=+0.619194388,LastTimestamp:2025-12-04 11:45:46.434524851 +0000 UTC m=+0.619194388,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:cert-recovery-controller,ReportingInstance:,}" 2025-12-04T11:45:57.910095767Z E1204 11:45:57.910014 1 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://localhost:6443/api/v1/namespaces/openshift-kube-controller-manager/events\": tls: failed to verify certificate: x509: certificate is valid for kubernetes, kubernetes.default, kubernetes.default.svc, kubernetes.default.svc.cluster.local, openshift, openshift.default, openshift.default.svc, openshift.default.svc.cluster.local, not localhost-recovery" event="&Event{ObjectMeta:{openshift-kube-controller-manager.187e00959585f6b3 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Namespace,Namespace:openshift-kube-controller-manager,Name:openshift-kube-controller-manager,UID:,APIVersion:v1,ResourceVersion:,FieldPath:,},Reason:ControlPlaneTopology,Message:unable to get control plane topology, using HA cluster values for leader election: Get \"https://localhost:6443/apis/config.openshift.io/v1/infrastructures/cluster\": tls: failed to verify certificate: x509: certificate is valid for kubernetes, kubernetes.default, kubernetes.default.svc, kubernetes.default.svc.cluster.local, openshift, openshift.default, openshift.default.svc, openshift.default.svc.cluster.local, not localhost-recovery,Source:EventSource{Component:cert-recovery-controller,Host:,},FirstTimestamp:2025-12-04 11:45:46.434524851 +0000 UTC m=+0.619194388,LastTimestamp:2025-12-04 11:45:46.434524851 +0000 UTC m=+0.619194388,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:cert-recovery-controller,ReportingInstance:,}" 2025-12-04T11:46:07.916721855Z E1204 11:46:07.916640 1 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://localhost:6443/api/v1/namespaces/openshift-kube-controller-manager/events\": tls: failed to verify certificate: x509: certificate is valid for kubernetes, kubernetes.default, kubernetes.default.svc, kubernetes.default.svc.cluster.local, openshift, openshift.default, openshift.default.svc, openshift.default.svc.cluster.local, not localhost-recovery" event="&Event{ObjectMeta:{openshift-kube-controller-manager.187e00959585f6b3 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Namespace,Namespace:openshift-kube-controller-manager,Name:openshift-kube-controller-manager,UID:,APIVersion:v1,ResourceVersion:,FieldPath:,},Reason:ControlPlaneTopology,Message:unable to get control plane topology, using HA cluster values for leader election: Get \"https://localhost:6443/apis/config.openshift.io/v1/infrastructures/cluster\": tls: failed to verify certificate: x509: certificate is valid for kubernetes, kubernetes.default, kubernetes.default.svc, kubernetes.default.svc.cluster.local, openshift, openshift.default, openshift.default.svc, openshift.default.svc.cluster.local, not localhost-recovery,Source:EventSource{Component:cert-recovery-controller,Host:,},FirstTimestamp:2025-12-04 11:45:46.434524851 +0000 UTC m=+0.619194388,LastTimestamp:2025-12-04 11:45:46.434524851 +0000 UTC m=+0.619194388,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:cert-recovery-controller,ReportingInstance:,}" 2025-12-04T11:46:17.922231481Z E1204 11:46:17.922130 1 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://localhost:6443/api/v1/namespaces/openshift-kube-controller-manager/events\": tls: failed to verify certificate: x509: certificate is valid for kubernetes, kubernetes.default, kubernetes.default.svc, kubernetes.default.svc.cluster.local, openshift, openshift.default, openshift.default.svc, openshift.default.svc.cluster.local, not localhost-recovery" event="&Event{ObjectMeta:{openshift-kube-controller-manager.187e00959585f6b3 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Namespace,Namespace:openshift-kube-controller-manager,Name:openshift-kube-controller-manager,UID:,APIVersion:v1,ResourceVersion:,FieldPath:,},Reason:ControlPlaneTopology,Message:unable to get control plane topology, using HA cluster values for leader election: Get \"https://localhost:6443/apis/config.openshift.io/v1/infrastructures/cluster\": tls: failed to verify certificate: x509: certificate is valid for kubernetes, kubernetes.default, kubernetes.default.svc, kubernetes.default.svc.cluster.local, openshift, openshift.default, openshift.default.svc, openshift.default.svc.cluster.local, not localhost-recovery,Source:EventSource{Component:cert-recovery-controller,Host:,},FirstTimestamp:2025-12-04 11:45:46.434524851 +0000 UTC m=+0.619194388,LastTimestamp:2025-12-04 11:45:46.434524851 +0000 UTC m=+0.619194388,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:cert-recovery-controller,ReportingInstance:,}" 2025-12-04T11:46:27.927089015Z E1204 11:46:27.926984 1 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://localhost:6443/api/v1/namespaces/openshift-kube-controller-manager/events\": tls: failed to verify certificate: x509: certificate is valid for kubernetes, kubernetes.default, kubernetes.default.svc, kubernetes.default.svc.cluster.local, openshift, openshift.default, openshift.default.svc, openshift.default.svc.cluster.local, not localhost-recovery" event="&Event{ObjectMeta:{openshift-kube-controller-manager.187e00959585f6b3 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Namespace,Namespace:openshift-kube-controller-manager,Name:openshift-kube-controller-manager,UID:,APIVersion:v1,ResourceVersion:,FieldPath:,},Reason:ControlPlaneTopology,Message:unable to get control plane topology, using HA cluster values for leader election: Get \"https://localhost:6443/apis/config.openshift.io/v1/infrastructures/cluster\": tls: failed to verify certificate: x509: certificate is valid for kubernetes, kubernetes.default, kubernetes.default.svc, kubernetes.default.svc.cluster.local, openshift, openshift.default, openshift.default.svc, openshift.default.svc.cluster.local, not localhost-recovery,Source:EventSource{Component:cert-recovery-controller,Host:,},FirstTimestamp:2025-12-04 11:45:46.434524851 +0000 UTC m=+0.619194388,LastTimestamp:2025-12-04 11:45:46.434524851 +0000 UTC m=+0.619194388,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:cert-recovery-controller,ReportingInstance:,}" 2025-12-04T11:46:37.932499267Z E1204 11:46:37.932403 1 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://localhost:6443/api/v1/namespaces/openshift-kube-controller-manager/events\": tls: failed to verify certificate: x509: certificate is valid for kubernetes, kubernetes.default, kubernetes.default.svc, kubernetes.default.svc.cluster.local, openshift, openshift.default, openshift.default.svc, openshift.default.svc.cluster.local, not localhost-recovery" event="&Event{ObjectMeta:{openshift-kube-controller-manager.187e00959585f6b3 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Namespace,Namespace:openshift-kube-controller-manager,Name:openshift-kube-controller-manager,UID:,APIVersion:v1,ResourceVersion:,FieldPath:,},Reason:ControlPlaneTopology,Message:unable to get control plane topology, using HA cluster values for leader election: Get \"https://localhost:6443/apis/config.openshift.io/v1/infrastructures/cluster\": tls: failed to verify certificate: x509: certificate is valid for kubernetes, kubernetes.default, kubernetes.default.svc, kubernetes.default.svc.cluster.local, openshift, openshift.default, openshift.default.svc, openshift.default.svc.cluster.local, not localhost-recovery,Source:EventSource{Component:cert-recovery-controller,Host:,},FirstTimestamp:2025-12-04 11:45:46.434524851 +0000 UTC m=+0.619194388,LastTimestamp:2025-12-04 11:45:46.434524851 +0000 UTC m=+0.619194388,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:cert-recovery-controller,ReportingInstance:,}" 2025-12-04T11:46:40.398255372Z E1204 11:46:40.398082 1 leaderelection.go:436] error retrieving resource lock openshift-kube-controller-manager/cert-recovery-controller-lock: Get "https://localhost:6443/apis/coordination.k8s.io/v1/namespaces/openshift-kube-controller-manager/leases/cert-recovery-controller-lock?timeout=1m47s": tls: failed to verify certificate: x509: certificate is valid for kubernetes, kubernetes.default, kubernetes.default.svc, kubernetes.default.svc.cluster.local, openshift, openshift.default, openshift.default.svc, openshift.default.svc.cluster.local, not localhost-recovery 2025-12-04T11:46:47.938082456Z E1204 11:46:47.938001 1 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://localhost:6443/api/v1/namespaces/openshift-kube-controller-manager/events\": tls: failed to verify certificate: x509: certificate is valid for kubernetes, kubernetes.default, kubernetes.default.svc, kubernetes.default.svc.cluster.local, openshift, openshift.default, openshift.default.svc, openshift.default.svc.cluster.local, not localhost-recovery" event="&Event{ObjectMeta:{openshift-kube-controller-manager.187e00959585f6b3 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Namespace,Namespace:openshift-kube-controller-manager,Name:openshift-kube-controller-manager,UID:,APIVersion:v1,ResourceVersion:,FieldPath:,},Reason:ControlPlaneTopology,Message:unable to get control plane topology, using HA cluster values for leader election: Get \"https://localhost:6443/apis/config.openshift.io/v1/infrastructures/cluster\": tls: failed to verify certificate: x509: certificate is valid for kubernetes, kubernetes.default, kubernetes.default.svc, kubernetes.default.svc.cluster.local, openshift, openshift.default, openshift.default.svc, openshift.default.svc.cluster.local, not localhost-recovery,Source:EventSource{Component:cert-recovery-controller,Host:,},FirstTimestamp:2025-12-04 11:45:46.434524851 +0000 UTC m=+0.619194388,LastTimestamp:2025-12-04 11:45:46.434524851 +0000 UTC m=+0.619194388,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:cert-recovery-controller,ReportingInstance:,}" 2025-12-04T11:46:57.944429379Z E1204 11:46:57.944327 1 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://localhost:6443/api/v1/namespaces/openshift-kube-controller-manager/events\": tls: failed to verify certificate: x509: certificate is valid for kubernetes, kubernetes.default, kubernetes.default.svc, kubernetes.default.svc.cluster.local, openshift, openshift.default, openshift.default.svc, openshift.default.svc.cluster.local, not localhost-recovery" event="&Event{ObjectMeta:{openshift-kube-controller-manager.187e00959585f6b3 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Namespace,Namespace:openshift-kube-controller-manager,Name:openshift-kube-controller-manager,UID:,APIVersion:v1,ResourceVersion:,FieldPath:,},Reason:ControlPlaneTopology,Message:unable to get control plane topology, using HA cluster values for leader election: Get \"https://localhost:6443/apis/config.openshift.io/v1/infrastructures/cluster\": tls: failed to verify certificate: x509: certificate is valid for kubernetes, kubernetes.default, kubernetes.default.svc, kubernetes.default.svc.cluster.local, openshift, openshift.default, openshift.default.svc, openshift.default.svc.cluster.local, not localhost-recovery,Source:EventSource{Component:cert-recovery-controller,Host:,},FirstTimestamp:2025-12-04 11:45:46.434524851 +0000 UTC m=+0.619194388,LastTimestamp:2025-12-04 11:45:46.434524851 +0000 UTC m=+0.619194388,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:cert-recovery-controller,ReportingInstance:,}" 2025-12-04T11:47:07.950961847Z E1204 11:47:07.950870 1 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://localhost:6443/api/v1/namespaces/openshift-kube-controller-manager/events\": tls: failed to verify certificate: x509: certificate is valid for kubernetes, kubernetes.default, kubernetes.default.svc, kubernetes.default.svc.cluster.local, openshift, openshift.default, openshift.default.svc, openshift.default.svc.cluster.local, not localhost-recovery" event="&Event{ObjectMeta:{openshift-kube-controller-manager.187e00959585f6b3 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Namespace,Namespace:openshift-kube-controller-manager,Name:openshift-kube-controller-manager,UID:,APIVersion:v1,ResourceVersion:,FieldPath:,},Reason:ControlPlaneTopology,Message:unable to get control plane topology, using HA cluster values for leader election: Get \"https://localhost:6443/apis/config.openshift.io/v1/infrastructures/cluster\": tls: failed to verify certificate: x509: certificate is valid for kubernetes, kubernetes.default, kubernetes.default.svc, kubernetes.default.svc.cluster.local, openshift, openshift.default, openshift.default.svc, openshift.default.svc.cluster.local, not localhost-recovery,Source:EventSource{Component:cert-recovery-controller,Host:,},FirstTimestamp:2025-12-04 11:45:46.434524851 +0000 UTC m=+0.619194388,LastTimestamp:2025-12-04 11:45:46.434524851 +0000 UTC m=+0.619194388,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:cert-recovery-controller,ReportingInstance:,}" 2025-12-04T11:47:17.955402666Z E1204 11:47:17.955308 1 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://localhost:6443/api/v1/namespaces/openshift-kube-controller-manager/events\": tls: failed to verify certificate: x509: certificate is valid for kubernetes, kubernetes.default, kubernetes.default.svc, kubernetes.default.svc.cluster.local, openshift, openshift.default, openshift.default.svc, openshift.default.svc.cluster.local, not localhost-recovery" event="&Event{ObjectMeta:{openshift-kube-controller-manager.187e00959585f6b3 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Namespace,Namespace:openshift-kube-controller-manager,Name:openshift-kube-controller-manager,UID:,APIVersion:v1,ResourceVersion:,FieldPath:,},Reason:ControlPlaneTopology,Message:unable to get control plane topology, using HA cluster values for leader election: Get \"https://localhost:6443/apis/config.openshift.io/v1/infrastructures/cluster\": tls: failed to verify certificate: x509: certificate is valid for kubernetes, kubernetes.default, kubernetes.default.svc, kubernetes.default.svc.cluster.local, openshift, openshift.default, openshift.default.svc, openshift.default.svc.cluster.local, not localhost-recovery,Source:EventSource{Component:cert-recovery-controller,Host:,},FirstTimestamp:2025-12-04 11:45:46.434524851 +0000 UTC m=+0.619194388,LastTimestamp:2025-12-04 11:45:46.434524851 +0000 UTC m=+0.619194388,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:cert-recovery-controller,ReportingInstance:,}" 2025-12-04T11:47:23.839311896Z E1204 11:47:23.839212 1 leaderelection.go:436] error retrieving resource lock openshift-kube-controller-manager/cert-recovery-controller-lock: Get "https://localhost:6443/apis/coordination.k8s.io/v1/namespaces/openshift-kube-controller-manager/leases/cert-recovery-controller-lock?timeout=1m47s": tls: failed to verify certificate: x509: certificate is valid for kubernetes, kubernetes.default, kubernetes.default.svc, kubernetes.default.svc.cluster.local, openshift, openshift.default, openshift.default.svc, openshift.default.svc.cluster.local, not localhost-recovery 2025-12-04T11:47:27.960279017Z E1204 11:47:27.960213 1 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://localhost:6443/api/v1/namespaces/openshift-kube-controller-manager/events\": tls: failed to verify certificate: x509: certificate is valid for kubernetes, kubernetes.default, kubernetes.default.svc, kubernetes.default.svc.cluster.local, openshift, openshift.default, openshift.default.svc, openshift.default.svc.cluster.local, not localhost-recovery" event="&Event{ObjectMeta:{openshift-kube-controller-manager.187e00959585f6b3 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Namespace,Namespace:openshift-kube-controller-manager,Name:openshift-kube-controller-manager,UID:,APIVersion:v1,ResourceVersion:,FieldPath:,},Reason:ControlPlaneTopology,Message:unable to get control plane topology, using HA cluster values for leader election: Get \"https://localhost:6443/apis/config.openshift.io/v1/infrastructures/cluster\": tls: failed to verify certificate: x509: certificate is valid for kubernetes, kubernetes.default, kubernetes.default.svc, kubernetes.default.svc.cluster.local, openshift, openshift.default, openshift.default.svc, openshift.default.svc.cluster.local, not localhost-recovery,Source:EventSource{Component:cert-recovery-controller,Host:,},FirstTimestamp:2025-12-04 11:45:46.434524851 +0000 UTC m=+0.619194388,LastTimestamp:2025-12-04 11:45:46.434524851 +0000 UTC m=+0.619194388,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:cert-recovery-controller,ReportingInstance:,}" 2025-12-04T11:47:27.960279017Z E1204 11:47:27.960267 1 event.go:307] "Unable to write event (retry limit exceeded!)" event="&Event{ObjectMeta:{openshift-kube-controller-manager.187e00959585f6b3 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Namespace,Namespace:openshift-kube-controller-manager,Name:openshift-kube-controller-manager,UID:,APIVersion:v1,ResourceVersion:,FieldPath:,},Reason:ControlPlaneTopology,Message:unable to get control plane topology, using HA cluster values for leader election: Get \"https://localhost:6443/apis/config.openshift.io/v1/infrastructures/cluster\": tls: failed to verify certificate: x509: certificate is valid for kubernetes, kubernetes.default, kubernetes.default.svc, kubernetes.default.svc.cluster.local, openshift, openshift.default, openshift.default.svc, openshift.default.svc.cluster.local, not localhost-recovery,Source:EventSource{Component:cert-recovery-controller,Host:,},FirstTimestamp:2025-12-04 11:45:46.434524851 +0000 UTC m=+0.619194388,LastTimestamp:2025-12-04 11:45:46.434524851 +0000 UTC m=+0.619194388,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:cert-recovery-controller,ReportingInstance:,}" 2025-12-04T11:48:16.262950196Z E1204 11:48:16.262844 1 leaderelection.go:436] error retrieving resource lock openshift-kube-controller-manager/cert-recovery-controller-lock: Get "https://localhost:6443/apis/coordination.k8s.io/v1/namespaces/openshift-kube-controller-manager/leases/cert-recovery-controller-lock?timeout=1m47s": tls: failed to verify certificate: x509: certificate is valid for kubernetes, kubernetes.default, kubernetes.default.svc, kubernetes.default.svc.cluster.local, openshift, openshift.default, openshift.default.svc, openshift.default.svc.cluster.local, not localhost-recovery 2025-12-04T11:49:12.111613381Z E1204 11:49:12.111503 1 leaderelection.go:436] error retrieving resource lock openshift-kube-controller-manager/cert-recovery-controller-lock: Get "https://localhost:6443/apis/coordination.k8s.io/v1/namespaces/openshift-kube-controller-manager/leases/cert-recovery-controller-lock?timeout=1m47s": tls: failed to verify certificate: x509: certificate is valid for kubernetes, kubernetes.default, kubernetes.default.svc, kubernetes.default.svc.cluster.local, openshift, openshift.default, openshift.default.svc, openshift.default.svc.cluster.local, not localhost-recovery 2025-12-04T11:49:49.053667287Z E1204 11:49:49.053579 1 leaderelection.go:436] error retrieving resource lock openshift-kube-controller-manager/cert-recovery-controller-lock: Get "https://localhost:6443/apis/coordination.k8s.io/v1/namespaces/openshift-kube-controller-manager/leases/cert-recovery-controller-lock?timeout=1m47s": tls: failed to verify certificate: x509: certificate is valid for kubernetes, kubernetes.default, kubernetes.default.svc, kubernetes.default.svc.cluster.local, openshift, openshift.default, openshift.default.svc, openshift.default.svc.cluster.local, not localhost-recovery 2025-12-04T11:50:41.302166997Z E1204 11:50:41.302064 1 leaderelection.go:436] error retrieving resource lock openshift-kube-controller-manager/cert-recovery-controller-lock: Get "https://localhost:6443/apis/coordination.k8s.io/v1/namespaces/openshift-kube-controller-manager/leases/cert-recovery-controller-lock?timeout=1m47s": tls: failed to verify certificate: x509: certificate is valid for kubernetes, kubernetes.default, kubernetes.default.svc, kubernetes.default.svc.cluster.local, openshift, openshift.default, openshift.default.svc, openshift.default.svc.cluster.local, not localhost-recovery 2025-12-04T11:51:14.022442771Z E1204 11:51:14.022344 1 leaderelection.go:436] error retrieving resource lock openshift-kube-controller-manager/cert-recovery-controller-lock: Get "https://localhost:6443/apis/coordination.k8s.io/v1/namespaces/openshift-kube-controller-manager/leases/cert-recovery-controller-lock?timeout=1m47s": tls: failed to verify certificate: x509: certificate is valid for kubernetes, kubernetes.default, kubernetes.default.svc, kubernetes.default.svc.cluster.local, openshift, openshift.default, openshift.default.svc, openshift.default.svc.cluster.local, not localhost-recovery 2025-12-04T11:51:56.439550012Z E1204 11:51:56.439448 1 leaderelection.go:436] error retrieving resource lock openshift-kube-controller-manager/cert-recovery-controller-lock: Get "https://localhost:6443/apis/coordination.k8s.io/v1/namespaces/openshift-kube-controller-manager/leases/cert-recovery-controller-lock?timeout=1m47s": tls: failed to verify certificate: x509: certificate is valid for kubernetes, kubernetes.default, kubernetes.default.svc, kubernetes.default.svc.cluster.local, openshift, openshift.default, openshift.default.svc, openshift.default.svc.cluster.local, not localhost-recovery 2025-12-04T11:52:51.164729874Z E1204 11:52:51.164621 1 leaderelection.go:436] error retrieving resource lock openshift-kube-controller-manager/cert-recovery-controller-lock: Get "https://localhost:6443/apis/coordination.k8s.io/v1/namespaces/openshift-kube-controller-manager/leases/cert-recovery-controller-lock?timeout=1m47s": tls: failed to verify certificate: x509: certificate is valid for kubernetes, kubernetes.default, kubernetes.default.svc, kubernetes.default.svc.cluster.local, openshift, openshift.default, openshift.default.svc, openshift.default.svc.cluster.local, not localhost-recovery 2025-12-04T11:53:37.168238105Z E1204 11:53:37.168173 1 leaderelection.go:436] error retrieving resource lock openshift-kube-controller-manager/cert-recovery-controller-lock: Get "https://localhost:6443/apis/coordination.k8s.io/v1/namespaces/openshift-kube-controller-manager/leases/cert-recovery-controller-lock?timeout=1m47s": tls: failed to verify certificate: x509: certificate is valid for kubernetes, kubernetes.default, kubernetes.default.svc, kubernetes.default.svc.cluster.local, openshift, openshift.default, openshift.default.svc, openshift.default.svc.cluster.local, not localhost-recovery 2025-12-04T11:54:07.182285007Z E1204 11:54:07.182178 1 leaderelection.go:436] error retrieving resource lock openshift-kube-controller-manager/cert-recovery-controller-lock: Get "https://localhost:6443/apis/coordination.k8s.io/v1/namespaces/openshift-kube-controller-manager/leases/cert-recovery-controller-lock?timeout=1m47s": tls: failed to verify certificate: x509: certificate is valid for kubernetes, kubernetes.default, kubernetes.default.svc, kubernetes.default.svc.cluster.local, openshift, openshift.default, openshift.default.svc, openshift.default.svc.cluster.local, not localhost-recovery 2025-12-04T11:54:41.148524661Z E1204 11:54:41.148428 1 leaderelection.go:436] error retrieving resource lock openshift-kube-controller-manager/cert-recovery-controller-lock: Get "https://localhost:6443/apis/coordination.k8s.io/v1/namespaces/openshift-kube-controller-manager/leases/cert-recovery-controller-lock?timeout=1m47s": tls: failed to verify certificate: x509: certificate is valid for kubernetes, kubernetes.default, kubernetes.default.svc, kubernetes.default.svc.cluster.local, openshift, openshift.default, openshift.default.svc, openshift.default.svc.cluster.local, not localhost-recovery 2025-12-04T11:55:09.823396041Z E1204 11:55:09.823319 1 leaderelection.go:436] error retrieving resource lock openshift-kube-controller-manager/cert-recovery-controller-lock: Get "https://localhost:6443/apis/coordination.k8s.io/v1/namespaces/openshift-kube-controller-manager/leases/cert-recovery-controller-lock?timeout=1m47s": tls: failed to verify certificate: x509: certificate is valid for kubernetes, kubernetes.default, kubernetes.default.svc, kubernetes.default.svc.cluster.local, openshift, openshift.default, openshift.default.svc, openshift.default.svc.cluster.local, not localhost-recovery 2025-12-04T11:56:07.011089524Z E1204 11:56:07.010991 1 leaderelection.go:436] error retrieving resource lock openshift-kube-controller-manager/cert-recovery-controller-lock: Get "https://localhost:6443/apis/coordination.k8s.io/v1/namespaces/openshift-kube-controller-manager/leases/cert-recovery-controller-lock?timeout=1m47s": tls: failed to verify certificate: x509: certificate is valid for kubernetes, kubernetes.default, kubernetes.default.svc, kubernetes.default.svc.cluster.local, openshift, openshift.default, openshift.default.svc, openshift.default.svc.cluster.local, not localhost-recovery 2025-12-04T11:56:42.356464904Z E1204 11:56:42.356372 1 leaderelection.go:436] error retrieving resource lock openshift-kube-controller-manager/cert-recovery-controller-lock: Get "https://localhost:6443/apis/coordination.k8s.io/v1/namespaces/openshift-kube-controller-manager/leases/cert-recovery-controller-lock?timeout=1m47s": tls: failed to verify certificate: x509: certificate is valid for kubernetes, kubernetes.default, kubernetes.default.svc, kubernetes.default.svc.cluster.local, openshift, openshift.default, openshift.default.svc, openshift.default.svc.cluster.local, not localhost-recovery 2025-12-04T11:57:38.463214122Z E1204 11:57:38.463065 1 leaderelection.go:436] error retrieving resource lock openshift-kube-controller-manager/cert-recovery-controller-lock: Get "https://localhost:6443/apis/coordination.k8s.io/v1/namespaces/openshift-kube-controller-manager/leases/cert-recovery-controller-lock?timeout=1m47s": tls: failed to verify certificate: x509: certificate is valid for kubernetes, kubernetes.default, kubernetes.default.svc, kubernetes.default.svc.cluster.local, openshift, openshift.default, openshift.default.svc, openshift.default.svc.cluster.local, not localhost-recovery 2025-12-04T11:58:30.587611684Z E1204 11:58:30.587499 1 leaderelection.go:436] error retrieving resource lock openshift-kube-controller-manager/cert-recovery-controller-lock: Get "https://localhost:6443/apis/coordination.k8s.io/v1/namespaces/openshift-kube-controller-manager/leases/cert-recovery-controller-lock?timeout=1m47s": tls: failed to verify certificate: x509: certificate is valid for kubernetes, kubernetes.default, kubernetes.default.svc, kubernetes.default.svc.cluster.local, openshift, openshift.default, openshift.default.svc, openshift.default.svc.cluster.local, not localhost-recovery 2025-12-04T11:59:14.553013139Z E1204 11:59:14.551887 1 leaderelection.go:436] error retrieving resource lock openshift-kube-controller-manager/cert-recovery-controller-lock: Get "https://localhost:6443/apis/coordination.k8s.io/v1/namespaces/openshift-kube-controller-manager/leases/cert-recovery-controller-lock?timeout=1m47s": tls: failed to verify certificate: x509: certificate is valid for kubernetes, kubernetes.default, kubernetes.default.svc, kubernetes.default.svc.cluster.local, openshift, openshift.default, openshift.default.svc, openshift.default.svc.cluster.local, not localhost-recovery 2025-12-04T11:59:57.245511574Z E1204 11:59:57.245283 1 leaderelection.go:436] error retrieving resource lock openshift-kube-controller-manager/cert-recovery-controller-lock: Get "https://localhost:6443/apis/coordination.k8s.io/v1/namespaces/openshift-kube-controller-manager/leases/cert-recovery-controller-lock?timeout=1m47s": tls: failed to verify certificate: x509: certificate signed by unknown authority 2025-12-04T12:00:31.466891661Z E1204 12:00:31.466746 1 leaderelection.go:436] error retrieving resource lock openshift-kube-controller-manager/cert-recovery-controller-lock: Get "https://localhost:6443/apis/coordination.k8s.io/v1/namespaces/openshift-kube-controller-manager/leases/cert-recovery-controller-lock?timeout=1m47s": tls: failed to verify certificate: x509: certificate signed by unknown authority 2025-12-04T12:01:15.425774403Z E1204 12:01:15.425705 1 leaderelection.go:436] error retrieving resource lock openshift-kube-controller-manager/cert-recovery-controller-lock: Get "https://localhost:6443/apis/coordination.k8s.io/v1/namespaces/openshift-kube-controller-manager/leases/cert-recovery-controller-lock?timeout=1m47s": tls: failed to verify certificate: x509: certificate signed by unknown authority kube-controller-manager-cert-syncer_current.log0000640000000000000000000006773015114274105040272 0ustar0000000000000000config/pod/openshift-kube-controller-manager/logs/kube-controller-manager-master-02025-12-04T11:55:46.411123455Z I1204 11:55:46.410734 1 base_controller.go:76] Waiting for caches to sync for CertSyncController 2025-12-04T11:55:46.411324652Z I1204 11:55:46.410737 1 observer_polling.go:159] Starting file observer 2025-12-04T11:55:46.418289703Z W1204 11:55:46.418128 1 reflector.go:561] k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243: failed to list *v1.ConfigMap: Get "https://localhost:6443/api/v1/namespaces/openshift-kube-controller-manager/configmaps?limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate is valid for kubernetes, kubernetes.default, kubernetes.default.svc, kubernetes.default.svc.cluster.local, openshift, openshift.default, openshift.default.svc, openshift.default.svc.cluster.local, not localhost-recovery 2025-12-04T11:55:46.418336644Z E1204 11:55:46.418292 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243: Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: Get \"https://localhost:6443/api/v1/namespaces/openshift-kube-controller-manager/configmaps?limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate is valid for kubernetes, kubernetes.default, kubernetes.default.svc, kubernetes.default.svc.cluster.local, openshift, openshift.default, openshift.default.svc, openshift.default.svc.cluster.local, not localhost-recovery" 2025-12-04T11:55:46.418774567Z W1204 11:55:46.418558 1 reflector.go:561] k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243: failed to list *v1.Secret: Get "https://localhost:6443/api/v1/namespaces/openshift-kube-controller-manager/secrets?limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate is valid for kubernetes, kubernetes.default, kubernetes.default.svc, kubernetes.default.svc.cluster.local, openshift, openshift.default, openshift.default.svc, openshift.default.svc.cluster.local, not localhost-recovery 2025-12-04T11:55:46.418858750Z E1204 11:55:46.418821 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243: Failed to watch *v1.Secret: failed to list *v1.Secret: Get \"https://localhost:6443/api/v1/namespaces/openshift-kube-controller-manager/secrets?limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate is valid for kubernetes, kubernetes.default, kubernetes.default.svc, kubernetes.default.svc.cluster.local, openshift, openshift.default, openshift.default.svc, openshift.default.svc.cluster.local, not localhost-recovery" 2025-12-04T11:55:47.248166965Z W1204 11:55:47.247991 1 reflector.go:561] k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243: failed to list *v1.Secret: Get "https://localhost:6443/api/v1/namespaces/openshift-kube-controller-manager/secrets?limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate is valid for kubernetes, kubernetes.default, kubernetes.default.svc, kubernetes.default.svc.cluster.local, openshift, openshift.default, openshift.default.svc, openshift.default.svc.cluster.local, not localhost-recovery 2025-12-04T11:55:47.248166965Z E1204 11:55:47.248088 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243: Failed to watch *v1.Secret: failed to list *v1.Secret: Get \"https://localhost:6443/api/v1/namespaces/openshift-kube-controller-manager/secrets?limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate is valid for kubernetes, kubernetes.default, kubernetes.default.svc, kubernetes.default.svc.cluster.local, openshift, openshift.default, openshift.default.svc, openshift.default.svc.cluster.local, not localhost-recovery" 2025-12-04T11:55:47.736870095Z W1204 11:55:47.735204 1 reflector.go:561] k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243: failed to list *v1.ConfigMap: Get "https://localhost:6443/api/v1/namespaces/openshift-kube-controller-manager/configmaps?limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate is valid for kubernetes, kubernetes.default, kubernetes.default.svc, kubernetes.default.svc.cluster.local, openshift, openshift.default, openshift.default.svc, openshift.default.svc.cluster.local, not localhost-recovery 2025-12-04T11:55:47.736870095Z E1204 11:55:47.735282 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243: Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: Get \"https://localhost:6443/api/v1/namespaces/openshift-kube-controller-manager/configmaps?limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate is valid for kubernetes, kubernetes.default, kubernetes.default.svc, kubernetes.default.svc.cluster.local, openshift, openshift.default, openshift.default.svc, openshift.default.svc.cluster.local, not localhost-recovery" 2025-12-04T11:55:49.651608232Z W1204 11:55:49.651484 1 reflector.go:561] k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243: failed to list *v1.Secret: Get "https://localhost:6443/api/v1/namespaces/openshift-kube-controller-manager/secrets?limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate is valid for kubernetes, kubernetes.default, kubernetes.default.svc, kubernetes.default.svc.cluster.local, openshift, openshift.default, openshift.default.svc, openshift.default.svc.cluster.local, not localhost-recovery 2025-12-04T11:55:49.651608232Z E1204 11:55:49.651562 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243: Failed to watch *v1.Secret: failed to list *v1.Secret: Get \"https://localhost:6443/api/v1/namespaces/openshift-kube-controller-manager/secrets?limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate is valid for kubernetes, kubernetes.default, kubernetes.default.svc, kubernetes.default.svc.cluster.local, openshift, openshift.default, openshift.default.svc, openshift.default.svc.cluster.local, not localhost-recovery" 2025-12-04T11:55:50.401928638Z W1204 11:55:50.401827 1 reflector.go:561] k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243: failed to list *v1.ConfigMap: Get "https://localhost:6443/api/v1/namespaces/openshift-kube-controller-manager/configmaps?limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate is valid for kubernetes, kubernetes.default, kubernetes.default.svc, kubernetes.default.svc.cluster.local, openshift, openshift.default, openshift.default.svc, openshift.default.svc.cluster.local, not localhost-recovery 2025-12-04T11:55:50.401928638Z E1204 11:55:50.401900 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243: Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: Get \"https://localhost:6443/api/v1/namespaces/openshift-kube-controller-manager/configmaps?limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate is valid for kubernetes, kubernetes.default, kubernetes.default.svc, kubernetes.default.svc.cluster.local, openshift, openshift.default, openshift.default.svc, openshift.default.svc.cluster.local, not localhost-recovery" 2025-12-04T11:55:53.457610895Z W1204 11:55:53.457487 1 reflector.go:561] k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243: failed to list *v1.Secret: Get "https://localhost:6443/api/v1/namespaces/openshift-kube-controller-manager/secrets?limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate is valid for kubernetes, kubernetes.default, kubernetes.default.svc, kubernetes.default.svc.cluster.local, openshift, openshift.default, openshift.default.svc, openshift.default.svc.cluster.local, not localhost-recovery 2025-12-04T11:55:53.457610895Z E1204 11:55:53.457583 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243: Failed to watch *v1.Secret: failed to list *v1.Secret: Get \"https://localhost:6443/api/v1/namespaces/openshift-kube-controller-manager/secrets?limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate is valid for kubernetes, kubernetes.default, kubernetes.default.svc, kubernetes.default.svc.cluster.local, openshift, openshift.default, openshift.default.svc, openshift.default.svc.cluster.local, not localhost-recovery" 2025-12-04T11:55:55.014982720Z W1204 11:55:55.014891 1 reflector.go:561] k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243: failed to list *v1.ConfigMap: Get "https://localhost:6443/api/v1/namespaces/openshift-kube-controller-manager/configmaps?limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate is valid for kubernetes, kubernetes.default, kubernetes.default.svc, kubernetes.default.svc.cluster.local, openshift, openshift.default, openshift.default.svc, openshift.default.svc.cluster.local, not localhost-recovery 2025-12-04T11:55:55.014982720Z E1204 11:55:55.014955 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243: Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: Get \"https://localhost:6443/api/v1/namespaces/openshift-kube-controller-manager/configmaps?limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate is valid for kubernetes, kubernetes.default, kubernetes.default.svc, kubernetes.default.svc.cluster.local, openshift, openshift.default, openshift.default.svc, openshift.default.svc.cluster.local, not localhost-recovery" 2025-12-04T11:56:03.196009588Z W1204 11:56:03.195939 1 reflector.go:561] k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243: failed to list *v1.ConfigMap: Get "https://localhost:6443/api/v1/namespaces/openshift-kube-controller-manager/configmaps?limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate is valid for kubernetes, kubernetes.default, kubernetes.default.svc, kubernetes.default.svc.cluster.local, openshift, openshift.default, openshift.default.svc, openshift.default.svc.cluster.local, not localhost-recovery 2025-12-04T11:56:03.196063689Z E1204 11:56:03.196038 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243: Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: Get \"https://localhost:6443/api/v1/namespaces/openshift-kube-controller-manager/configmaps?limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate is valid for kubernetes, kubernetes.default, kubernetes.default.svc, kubernetes.default.svc.cluster.local, openshift, openshift.default, openshift.default.svc, openshift.default.svc.cluster.local, not localhost-recovery" 2025-12-04T11:56:04.139127838Z W1204 11:56:04.139035 1 reflector.go:561] k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243: failed to list *v1.Secret: Get "https://localhost:6443/api/v1/namespaces/openshift-kube-controller-manager/secrets?limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate is valid for kubernetes, kubernetes.default, kubernetes.default.svc, kubernetes.default.svc.cluster.local, openshift, openshift.default, openshift.default.svc, openshift.default.svc.cluster.local, not localhost-recovery 2025-12-04T11:56:04.139127838Z E1204 11:56:04.139116 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243: Failed to watch *v1.Secret: failed to list *v1.Secret: Get \"https://localhost:6443/api/v1/namespaces/openshift-kube-controller-manager/secrets?limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate is valid for kubernetes, kubernetes.default, kubernetes.default.svc, kubernetes.default.svc.cluster.local, openshift, openshift.default, openshift.default.svc, openshift.default.svc.cluster.local, not localhost-recovery" 2025-12-04T11:56:23.856938320Z W1204 11:56:23.856846 1 reflector.go:561] k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243: failed to list *v1.Secret: Get "https://localhost:6443/api/v1/namespaces/openshift-kube-controller-manager/secrets?limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate is valid for kubernetes, kubernetes.default, kubernetes.default.svc, kubernetes.default.svc.cluster.local, openshift, openshift.default, openshift.default.svc, openshift.default.svc.cluster.local, not localhost-recovery 2025-12-04T11:56:23.856938320Z E1204 11:56:23.856910 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243: Failed to watch *v1.Secret: failed to list *v1.Secret: Get \"https://localhost:6443/api/v1/namespaces/openshift-kube-controller-manager/secrets?limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate is valid for kubernetes, kubernetes.default, kubernetes.default.svc, kubernetes.default.svc.cluster.local, openshift, openshift.default, openshift.default.svc, openshift.default.svc.cluster.local, not localhost-recovery" 2025-12-04T11:56:25.097505824Z W1204 11:56:25.097398 1 reflector.go:561] k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243: failed to list *v1.ConfigMap: Get "https://localhost:6443/api/v1/namespaces/openshift-kube-controller-manager/configmaps?limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate is valid for kubernetes, kubernetes.default, kubernetes.default.svc, kubernetes.default.svc.cluster.local, openshift, openshift.default, openshift.default.svc, openshift.default.svc.cluster.local, not localhost-recovery 2025-12-04T11:56:25.097565736Z E1204 11:56:25.097508 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243: Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: Get \"https://localhost:6443/api/v1/namespaces/openshift-kube-controller-manager/configmaps?limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate is valid for kubernetes, kubernetes.default, kubernetes.default.svc, kubernetes.default.svc.cluster.local, openshift, openshift.default, openshift.default.svc, openshift.default.svc.cluster.local, not localhost-recovery" 2025-12-04T11:57:07.694991233Z W1204 11:57:07.694708 1 reflector.go:561] k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243: failed to list *v1.ConfigMap: Get "https://localhost:6443/api/v1/namespaces/openshift-kube-controller-manager/configmaps?limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate is valid for kubernetes, kubernetes.default, kubernetes.default.svc, kubernetes.default.svc.cluster.local, openshift, openshift.default, openshift.default.svc, openshift.default.svc.cluster.local, not localhost-recovery 2025-12-04T11:57:07.694991233Z E1204 11:57:07.694863 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243: Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: Get \"https://localhost:6443/api/v1/namespaces/openshift-kube-controller-manager/configmaps?limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate is valid for kubernetes, kubernetes.default, kubernetes.default.svc, kubernetes.default.svc.cluster.local, openshift, openshift.default, openshift.default.svc, openshift.default.svc.cluster.local, not localhost-recovery" 2025-12-04T11:57:07.711867294Z W1204 11:57:07.711694 1 reflector.go:561] k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243: failed to list *v1.Secret: Get "https://localhost:6443/api/v1/namespaces/openshift-kube-controller-manager/secrets?limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate is valid for kubernetes, kubernetes.default, kubernetes.default.svc, kubernetes.default.svc.cluster.local, openshift, openshift.default, openshift.default.svc, openshift.default.svc.cluster.local, not localhost-recovery 2025-12-04T11:57:07.711867294Z E1204 11:57:07.711847 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243: Failed to watch *v1.Secret: failed to list *v1.Secret: Get \"https://localhost:6443/api/v1/namespaces/openshift-kube-controller-manager/secrets?limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate is valid for kubernetes, kubernetes.default, kubernetes.default.svc, kubernetes.default.svc.cluster.local, openshift, openshift.default, openshift.default.svc, openshift.default.svc.cluster.local, not localhost-recovery" 2025-12-04T11:57:44.688679362Z W1204 11:57:44.688513 1 reflector.go:561] k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243: failed to list *v1.ConfigMap: Get "https://localhost:6443/api/v1/namespaces/openshift-kube-controller-manager/configmaps?limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate is valid for kubernetes, kubernetes.default, kubernetes.default.svc, kubernetes.default.svc.cluster.local, openshift, openshift.default, openshift.default.svc, openshift.default.svc.cluster.local, not localhost-recovery 2025-12-04T11:57:44.688679362Z E1204 11:57:44.688636 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243: Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: Get \"https://localhost:6443/api/v1/namespaces/openshift-kube-controller-manager/configmaps?limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate is valid for kubernetes, kubernetes.default, kubernetes.default.svc, kubernetes.default.svc.cluster.local, openshift, openshift.default, openshift.default.svc, openshift.default.svc.cluster.local, not localhost-recovery" 2025-12-04T11:57:48.559480363Z W1204 11:57:48.559386 1 reflector.go:561] k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243: failed to list *v1.Secret: Get "https://localhost:6443/api/v1/namespaces/openshift-kube-controller-manager/secrets?limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate is valid for kubernetes, kubernetes.default, kubernetes.default.svc, kubernetes.default.svc.cluster.local, openshift, openshift.default, openshift.default.svc, openshift.default.svc.cluster.local, not localhost-recovery 2025-12-04T11:57:48.559480363Z E1204 11:57:48.559464 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243: Failed to watch *v1.Secret: failed to list *v1.Secret: Get \"https://localhost:6443/api/v1/namespaces/openshift-kube-controller-manager/secrets?limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate is valid for kubernetes, kubernetes.default, kubernetes.default.svc, kubernetes.default.svc.cluster.local, openshift, openshift.default, openshift.default.svc, openshift.default.svc.cluster.local, not localhost-recovery" 2025-12-04T11:58:19.398102563Z W1204 11:58:19.397902 1 reflector.go:561] k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243: failed to list *v1.Secret: Get "https://localhost:6443/api/v1/namespaces/openshift-kube-controller-manager/secrets?limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate is valid for kubernetes, kubernetes.default, kubernetes.default.svc, kubernetes.default.svc.cluster.local, openshift, openshift.default, openshift.default.svc, openshift.default.svc.cluster.local, not localhost-recovery 2025-12-04T11:58:19.398102563Z E1204 11:58:19.398024 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243: Failed to watch *v1.Secret: failed to list *v1.Secret: Get \"https://localhost:6443/api/v1/namespaces/openshift-kube-controller-manager/secrets?limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate is valid for kubernetes, kubernetes.default, kubernetes.default.svc, kubernetes.default.svc.cluster.local, openshift, openshift.default, openshift.default.svc, openshift.default.svc.cluster.local, not localhost-recovery" 2025-12-04T11:58:37.799024374Z W1204 11:58:37.798929 1 reflector.go:561] k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243: failed to list *v1.ConfigMap: Get "https://localhost:6443/api/v1/namespaces/openshift-kube-controller-manager/configmaps?limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate is valid for kubernetes, kubernetes.default, kubernetes.default.svc, kubernetes.default.svc.cluster.local, openshift, openshift.default, openshift.default.svc, openshift.default.svc.cluster.local, not localhost-recovery 2025-12-04T11:58:37.799128697Z E1204 11:58:37.799114 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243: Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: Get \"https://localhost:6443/api/v1/namespaces/openshift-kube-controller-manager/configmaps?limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate is valid for kubernetes, kubernetes.default, kubernetes.default.svc, kubernetes.default.svc.cluster.local, openshift, openshift.default, openshift.default.svc, openshift.default.svc.cluster.local, not localhost-recovery" 2025-12-04T11:58:53.064615688Z W1204 11:58:53.064509 1 reflector.go:561] k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243: failed to list *v1.Secret: Get "https://localhost:6443/api/v1/namespaces/openshift-kube-controller-manager/secrets?limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate is valid for kubernetes, kubernetes.default, kubernetes.default.svc, kubernetes.default.svc.cluster.local, openshift, openshift.default, openshift.default.svc, openshift.default.svc.cluster.local, not localhost-recovery 2025-12-04T11:58:53.064615688Z E1204 11:58:53.064579 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243: Failed to watch *v1.Secret: failed to list *v1.Secret: Get \"https://localhost:6443/api/v1/namespaces/openshift-kube-controller-manager/secrets?limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate is valid for kubernetes, kubernetes.default, kubernetes.default.svc, kubernetes.default.svc.cluster.local, openshift, openshift.default, openshift.default.svc, openshift.default.svc.cluster.local, not localhost-recovery" 2025-12-04T11:59:29.260777767Z W1204 11:59:29.260646 1 reflector.go:561] k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243: failed to list *v1.Secret: Get "https://localhost:6443/api/v1/namespaces/openshift-kube-controller-manager/secrets?limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate signed by unknown authority 2025-12-04T11:59:29.260858020Z E1204 11:59:29.260785 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243: Failed to watch *v1.Secret: failed to list *v1.Secret: Get \"https://localhost:6443/api/v1/namespaces/openshift-kube-controller-manager/secrets?limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate signed by unknown authority" 2025-12-04T11:59:31.208218676Z W1204 11:59:31.208094 1 reflector.go:561] k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243: failed to list *v1.ConfigMap: Get "https://localhost:6443/api/v1/namespaces/openshift-kube-controller-manager/configmaps?limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate signed by unknown authority 2025-12-04T11:59:31.208218676Z E1204 11:59:31.208209 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243: Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: Get \"https://localhost:6443/api/v1/namespaces/openshift-kube-controller-manager/configmaps?limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate signed by unknown authority" 2025-12-04T12:00:25.692156930Z W1204 12:00:25.692044 1 reflector.go:561] k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243: failed to list *v1.Secret: Get "https://localhost:6443/api/v1/namespaces/openshift-kube-controller-manager/secrets?limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate signed by unknown authority 2025-12-04T12:00:25.692156930Z E1204 12:00:25.692140 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243: Failed to watch *v1.Secret: failed to list *v1.Secret: Get \"https://localhost:6443/api/v1/namespaces/openshift-kube-controller-manager/secrets?limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate signed by unknown authority" 2025-12-04T12:00:27.882851812Z W1204 12:00:27.882711 1 reflector.go:561] k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243: failed to list *v1.ConfigMap: Get "https://localhost:6443/api/v1/namespaces/openshift-kube-controller-manager/configmaps?limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate signed by unknown authority 2025-12-04T12:00:27.882851812Z E1204 12:00:27.882800 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243: Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: Get \"https://localhost:6443/api/v1/namespaces/openshift-kube-controller-manager/configmaps?limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate signed by unknown authority" 2025-12-04T12:01:07.598018971Z W1204 12:01:07.597879 1 reflector.go:561] k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243: failed to list *v1.Secret: Get "https://localhost:6443/api/v1/namespaces/openshift-kube-controller-manager/secrets?limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate signed by unknown authority 2025-12-04T12:01:07.598018971Z E1204 12:01:07.597998 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243: Failed to watch *v1.Secret: failed to list *v1.Secret: Get \"https://localhost:6443/api/v1/namespaces/openshift-kube-controller-manager/secrets?limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate signed by unknown authority" 2025-12-04T12:01:25.858885350Z W1204 12:01:25.858689 1 reflector.go:561] k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243: failed to list *v1.ConfigMap: Get "https://localhost:6443/api/v1/namespaces/openshift-kube-controller-manager/configmaps?limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate signed by unknown authority 2025-12-04T12:01:25.858885350Z E1204 12:01:25.858852 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243: Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: Get \"https://localhost:6443/api/v1/namespaces/openshift-kube-controller-manager/configmaps?limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate signed by unknown authority" 2025-12-04T12:01:56.703899577Z W1204 12:01:56.703729 1 reflector.go:561] k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243: failed to list *v1.ConfigMap: Get "https://localhost:6443/api/v1/namespaces/openshift-kube-controller-manager/configmaps?limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate signed by unknown authority 2025-12-04T12:01:56.703899577Z E1204 12:01:56.703870 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243: Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: Get \"https://localhost:6443/api/v1/namespaces/openshift-kube-controller-manager/configmaps?limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate signed by unknown authority" 2025-12-04T12:02:02.978827088Z W1204 12:02:02.978667 1 reflector.go:561] k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243: failed to list *v1.Secret: Get "https://localhost:6443/api/v1/namespaces/openshift-kube-controller-manager/secrets?limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate signed by unknown authority 2025-12-04T12:02:02.978827088Z E1204 12:02:02.978769 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243: Failed to watch *v1.Secret: failed to list *v1.Secret: Get \"https://localhost:6443/api/v1/namespaces/openshift-kube-controller-manager/secrets?limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate signed by unknown authority" cluster-policy-controller_current.log0000640000000000000000000000364015114274105036444 0ustar0000000000000000config/pod/openshift-kube-controller-manager/logs/kube-controller-manager-master-02025-12-04T12:01:28.412005903Z + timeout 3m /bin/bash -exuo pipefail -c 'while [ -n "$(ss -Htanop \( sport = 10357 \))" ]; do sleep 1; done' 2025-12-04T12:01:28.416748943Z ++ ss -Htanop '(' sport = 10357 ')' 2025-12-04T12:01:28.430581551Z + '[' -n '' ']' 2025-12-04T12:01:28.431564632Z + exec cluster-policy-controller start --config=/etc/kubernetes/static-pod-resources/configmaps/cluster-policy-controller-config/config.yaml --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager -v=2 2025-12-04T12:01:28.483117073Z I1204 12:01:28.482899 1 leaderelection.go:121] The leader election gives 4 retries and allows for 30s of clock skew. The kube-apiserver downtime tolerance is 78s. Worst non-graceful lease acquisition is 2m43s. Worst graceful lease acquisition is {26s}. 2025-12-04T12:01:28.483831916Z I1204 12:01:28.483760 1 observer_polling.go:159] Starting file observer 2025-12-04T12:01:28.484617021Z I1204 12:01:28.484564 1 builder.go:298] cluster-policy-controller version 4.18.0-202511181540.p2.g5fd8525.assembly.stream.el9-5fd8525-5fd852525909ce6eab52972ba9ce8fcf56528eb9 2025-12-04T12:01:28.485247621Z I1204 12:01:28.485189 1 dynamic_serving_content.go:116] "Loaded a new cert/key pair" name="serving-cert::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key" 2025-12-04T12:01:54.647387901Z I1204 12:01:54.647285 1 cmd.go:138] Received SIGTERM or SIGINT signal, shutting down controller. 2025-12-04T12:01:54.647438203Z F1204 12:01:54.647400 1 cmd.go:179] failed checking apiserver connectivity: Get "https://api-int.sno.openstack.lab:6443/apis/coordination.k8s.io/v1/namespaces/openshift-kube-controller-manager/leases/cluster-policy-controller-lock": tls: failed to verify certificate: x509: certificate signed by unknown authority kube-controller-manager_current.log0000640000000000000000000011044015114274105036021 0ustar0000000000000000config/pod/openshift-kube-controller-manager/logs/kube-controller-manager-master-02025-12-04T11:54:13.442210109Z E1204 11:54:13.442118 1 leaderelection.go:436] error retrieving resource lock kube-system/kube-controller-manager: Get "https://api-int.sno.openstack.lab:6443/apis/coordination.k8s.io/v1/namespaces/kube-system/leases/kube-controller-manager?timeout=6s": tls: failed to verify certificate: x509: certificate signed by unknown authority 2025-12-04T11:54:18.836547370Z E1204 11:54:18.836436 1 leaderelection.go:436] error retrieving resource lock kube-system/kube-controller-manager: Get "https://api-int.sno.openstack.lab:6443/apis/coordination.k8s.io/v1/namespaces/kube-system/leases/kube-controller-manager?timeout=6s": tls: failed to verify certificate: x509: certificate signed by unknown authority 2025-12-04T11:54:24.237579374Z E1204 11:54:24.237511 1 leaderelection.go:436] error retrieving resource lock kube-system/kube-controller-manager: Get "https://api-int.sno.openstack.lab:6443/apis/coordination.k8s.io/v1/namespaces/kube-system/leases/kube-controller-manager?timeout=6s": tls: failed to verify certificate: x509: certificate signed by unknown authority 2025-12-04T11:54:29.673599216Z E1204 11:54:29.673500 1 leaderelection.go:436] error retrieving resource lock kube-system/kube-controller-manager: Get "https://api-int.sno.openstack.lab:6443/apis/coordination.k8s.io/v1/namespaces/kube-system/leases/kube-controller-manager?timeout=6s": tls: failed to verify certificate: x509: certificate signed by unknown authority 2025-12-04T11:54:33.634677640Z E1204 11:54:33.634540 1 leaderelection.go:436] error retrieving resource lock kube-system/kube-controller-manager: Get "https://api-int.sno.openstack.lab:6443/apis/coordination.k8s.io/v1/namespaces/kube-system/leases/kube-controller-manager?timeout=6s": tls: failed to verify certificate: x509: certificate signed by unknown authority 2025-12-04T11:54:37.801298968Z E1204 11:54:37.801181 1 leaderelection.go:436] error retrieving resource lock kube-system/kube-controller-manager: Get "https://api-int.sno.openstack.lab:6443/apis/coordination.k8s.io/v1/namespaces/kube-system/leases/kube-controller-manager?timeout=6s": tls: failed to verify certificate: x509: certificate signed by unknown authority 2025-12-04T11:54:43.103662412Z E1204 11:54:43.103555 1 leaderelection.go:436] error retrieving resource lock kube-system/kube-controller-manager: Get "https://api-int.sno.openstack.lab:6443/apis/coordination.k8s.io/v1/namespaces/kube-system/leases/kube-controller-manager?timeout=6s": tls: failed to verify certificate: x509: certificate signed by unknown authority 2025-12-04T11:54:46.157950821Z E1204 11:54:46.157866 1 leaderelection.go:436] error retrieving resource lock kube-system/kube-controller-manager: Get "https://api-int.sno.openstack.lab:6443/apis/coordination.k8s.io/v1/namespaces/kube-system/leases/kube-controller-manager?timeout=6s": tls: failed to verify certificate: x509: certificate signed by unknown authority 2025-12-04T11:54:51.009130043Z E1204 11:54:51.009049 1 leaderelection.go:436] error retrieving resource lock kube-system/kube-controller-manager: Get "https://api-int.sno.openstack.lab:6443/apis/coordination.k8s.io/v1/namespaces/kube-system/leases/kube-controller-manager?timeout=6s": tls: failed to verify certificate: x509: certificate signed by unknown authority 2025-12-04T11:54:56.188533359Z E1204 11:54:56.188446 1 leaderelection.go:436] error retrieving resource lock kube-system/kube-controller-manager: Get "https://api-int.sno.openstack.lab:6443/apis/coordination.k8s.io/v1/namespaces/kube-system/leases/kube-controller-manager?timeout=6s": tls: failed to verify certificate: x509: certificate signed by unknown authority 2025-12-04T11:55:02.522147998Z E1204 11:55:02.522034 1 leaderelection.go:436] error retrieving resource lock kube-system/kube-controller-manager: Get "https://api-int.sno.openstack.lab:6443/apis/coordination.k8s.io/v1/namespaces/kube-system/leases/kube-controller-manager?timeout=6s": tls: failed to verify certificate: x509: certificate signed by unknown authority 2025-12-04T11:55:08.754863352Z E1204 11:55:08.754684 1 leaderelection.go:436] error retrieving resource lock kube-system/kube-controller-manager: Get "https://api-int.sno.openstack.lab:6443/apis/coordination.k8s.io/v1/namespaces/kube-system/leases/kube-controller-manager?timeout=6s": tls: failed to verify certificate: x509: certificate signed by unknown authority 2025-12-04T11:55:15.186288002Z E1204 11:55:15.186131 1 leaderelection.go:436] error retrieving resource lock kube-system/kube-controller-manager: Get "https://api-int.sno.openstack.lab:6443/apis/coordination.k8s.io/v1/namespaces/kube-system/leases/kube-controller-manager?timeout=6s": tls: failed to verify certificate: x509: certificate signed by unknown authority 2025-12-04T11:55:19.094757175Z E1204 11:55:19.094669 1 leaderelection.go:436] error retrieving resource lock kube-system/kube-controller-manager: Get "https://api-int.sno.openstack.lab:6443/apis/coordination.k8s.io/v1/namespaces/kube-system/leases/kube-controller-manager?timeout=6s": tls: failed to verify certificate: x509: certificate signed by unknown authority 2025-12-04T11:55:24.570053878Z E1204 11:55:24.569950 1 leaderelection.go:436] error retrieving resource lock kube-system/kube-controller-manager: Get "https://api-int.sno.openstack.lab:6443/apis/coordination.k8s.io/v1/namespaces/kube-system/leases/kube-controller-manager?timeout=6s": tls: failed to verify certificate: x509: certificate signed by unknown authority 2025-12-04T11:55:29.111468060Z E1204 11:55:29.111352 1 leaderelection.go:436] error retrieving resource lock kube-system/kube-controller-manager: Get "https://api-int.sno.openstack.lab:6443/apis/coordination.k8s.io/v1/namespaces/kube-system/leases/kube-controller-manager?timeout=6s": tls: failed to verify certificate: x509: certificate signed by unknown authority 2025-12-04T11:55:33.399358443Z E1204 11:55:33.399250 1 leaderelection.go:436] error retrieving resource lock kube-system/kube-controller-manager: Get "https://api-int.sno.openstack.lab:6443/apis/coordination.k8s.io/v1/namespaces/kube-system/leases/kube-controller-manager?timeout=6s": tls: failed to verify certificate: x509: certificate signed by unknown authority 2025-12-04T11:55:39.884252666Z E1204 11:55:39.884131 1 leaderelection.go:436] error retrieving resource lock kube-system/kube-controller-manager: Get "https://api-int.sno.openstack.lab:6443/apis/coordination.k8s.io/v1/namespaces/kube-system/leases/kube-controller-manager?timeout=6s": tls: failed to verify certificate: x509: certificate signed by unknown authority 2025-12-04T11:55:43.660249076Z E1204 11:55:43.660139 1 leaderelection.go:436] error retrieving resource lock kube-system/kube-controller-manager: Get "https://api-int.sno.openstack.lab:6443/apis/coordination.k8s.io/v1/namespaces/kube-system/leases/kube-controller-manager?timeout=6s": tls: failed to verify certificate: x509: certificate signed by unknown authority 2025-12-04T11:55:49.372425984Z E1204 11:55:49.372331 1 leaderelection.go:436] error retrieving resource lock kube-system/kube-controller-manager: Get "https://api-int.sno.openstack.lab:6443/apis/coordination.k8s.io/v1/namespaces/kube-system/leases/kube-controller-manager?timeout=6s": tls: failed to verify certificate: x509: certificate signed by unknown authority 2025-12-04T11:55:53.135394969Z E1204 11:55:53.135224 1 leaderelection.go:436] error retrieving resource lock kube-system/kube-controller-manager: Get "https://api-int.sno.openstack.lab:6443/apis/coordination.k8s.io/v1/namespaces/kube-system/leases/kube-controller-manager?timeout=6s": tls: failed to verify certificate: x509: certificate signed by unknown authority 2025-12-04T11:55:57.660516938Z E1204 11:55:57.660358 1 leaderelection.go:436] error retrieving resource lock kube-system/kube-controller-manager: Get "https://api-int.sno.openstack.lab:6443/apis/coordination.k8s.io/v1/namespaces/kube-system/leases/kube-controller-manager?timeout=6s": tls: failed to verify certificate: x509: certificate signed by unknown authority 2025-12-04T11:56:00.702326100Z E1204 11:56:00.702254 1 leaderelection.go:436] error retrieving resource lock kube-system/kube-controller-manager: Get "https://api-int.sno.openstack.lab:6443/apis/coordination.k8s.io/v1/namespaces/kube-system/leases/kube-controller-manager?timeout=6s": tls: failed to verify certificate: x509: certificate signed by unknown authority 2025-12-04T11:56:05.881852312Z E1204 11:56:05.881746 1 leaderelection.go:436] error retrieving resource lock kube-system/kube-controller-manager: Get "https://api-int.sno.openstack.lab:6443/apis/coordination.k8s.io/v1/namespaces/kube-system/leases/kube-controller-manager?timeout=6s": tls: failed to verify certificate: x509: certificate signed by unknown authority 2025-12-04T11:56:11.631869808Z E1204 11:56:11.630925 1 leaderelection.go:436] error retrieving resource lock kube-system/kube-controller-manager: Get "https://api-int.sno.openstack.lab:6443/apis/coordination.k8s.io/v1/namespaces/kube-system/leases/kube-controller-manager?timeout=6s": tls: failed to verify certificate: x509: certificate signed by unknown authority 2025-12-04T11:56:16.839027839Z E1204 11:56:16.838822 1 leaderelection.go:436] error retrieving resource lock kube-system/kube-controller-manager: Get "https://api-int.sno.openstack.lab:6443/apis/coordination.k8s.io/v1/namespaces/kube-system/leases/kube-controller-manager?timeout=6s": tls: failed to verify certificate: x509: certificate signed by unknown authority 2025-12-04T11:56:22.291989576Z E1204 11:56:22.291878 1 leaderelection.go:436] error retrieving resource lock kube-system/kube-controller-manager: Get "https://api-int.sno.openstack.lab:6443/apis/coordination.k8s.io/v1/namespaces/kube-system/leases/kube-controller-manager?timeout=6s": tls: failed to verify certificate: x509: certificate signed by unknown authority 2025-12-04T11:56:28.608156006Z E1204 11:56:28.608068 1 leaderelection.go:436] error retrieving resource lock kube-system/kube-controller-manager: Get "https://api-int.sno.openstack.lab:6443/apis/coordination.k8s.io/v1/namespaces/kube-system/leases/kube-controller-manager?timeout=6s": tls: failed to verify certificate: x509: certificate signed by unknown authority 2025-12-04T11:56:34.560434595Z E1204 11:56:34.560318 1 leaderelection.go:436] error retrieving resource lock kube-system/kube-controller-manager: Get "https://api-int.sno.openstack.lab:6443/apis/coordination.k8s.io/v1/namespaces/kube-system/leases/kube-controller-manager?timeout=6s": tls: failed to verify certificate: x509: certificate signed by unknown authority 2025-12-04T11:56:39.777007382Z E1204 11:56:39.776910 1 leaderelection.go:436] error retrieving resource lock kube-system/kube-controller-manager: Get "https://api-int.sno.openstack.lab:6443/apis/coordination.k8s.io/v1/namespaces/kube-system/leases/kube-controller-manager?timeout=6s": tls: failed to verify certificate: x509: certificate signed by unknown authority 2025-12-04T11:56:45.667999049Z E1204 11:56:45.667876 1 leaderelection.go:436] error retrieving resource lock kube-system/kube-controller-manager: Get "https://api-int.sno.openstack.lab:6443/apis/coordination.k8s.io/v1/namespaces/kube-system/leases/kube-controller-manager?timeout=6s": tls: failed to verify certificate: x509: certificate signed by unknown authority 2025-12-04T11:56:51.028467487Z E1204 11:56:51.028388 1 leaderelection.go:436] error retrieving resource lock kube-system/kube-controller-manager: Get "https://api-int.sno.openstack.lab:6443/apis/coordination.k8s.io/v1/namespaces/kube-system/leases/kube-controller-manager?timeout=6s": tls: failed to verify certificate: x509: certificate signed by unknown authority 2025-12-04T11:56:55.701799457Z E1204 11:56:55.701700 1 leaderelection.go:436] error retrieving resource lock kube-system/kube-controller-manager: Get "https://api-int.sno.openstack.lab:6443/apis/coordination.k8s.io/v1/namespaces/kube-system/leases/kube-controller-manager?timeout=6s": tls: failed to verify certificate: x509: certificate signed by unknown authority 2025-12-04T11:57:01.714351716Z E1204 11:57:01.714257 1 leaderelection.go:436] error retrieving resource lock kube-system/kube-controller-manager: Get "https://api-int.sno.openstack.lab:6443/apis/coordination.k8s.io/v1/namespaces/kube-system/leases/kube-controller-manager?timeout=6s": tls: failed to verify certificate: x509: certificate signed by unknown authority 2025-12-04T11:57:04.829236774Z E1204 11:57:04.828999 1 leaderelection.go:436] error retrieving resource lock kube-system/kube-controller-manager: Get "https://api-int.sno.openstack.lab:6443/apis/coordination.k8s.io/v1/namespaces/kube-system/leases/kube-controller-manager?timeout=6s": tls: failed to verify certificate: x509: certificate signed by unknown authority 2025-12-04T11:57:09.570153029Z E1204 11:57:09.570060 1 leaderelection.go:436] error retrieving resource lock kube-system/kube-controller-manager: Get "https://api-int.sno.openstack.lab:6443/apis/coordination.k8s.io/v1/namespaces/kube-system/leases/kube-controller-manager?timeout=6s": tls: failed to verify certificate: x509: certificate signed by unknown authority 2025-12-04T11:57:15.012433249Z E1204 11:57:15.012280 1 leaderelection.go:436] error retrieving resource lock kube-system/kube-controller-manager: Get "https://api-int.sno.openstack.lab:6443/apis/coordination.k8s.io/v1/namespaces/kube-system/leases/kube-controller-manager?timeout=6s": tls: failed to verify certificate: x509: certificate signed by unknown authority 2025-12-04T11:57:20.187259518Z E1204 11:57:20.187168 1 leaderelection.go:436] error retrieving resource lock kube-system/kube-controller-manager: Get "https://api-int.sno.openstack.lab:6443/apis/coordination.k8s.io/v1/namespaces/kube-system/leases/kube-controller-manager?timeout=6s": tls: failed to verify certificate: x509: certificate signed by unknown authority 2025-12-04T11:57:24.939642679Z E1204 11:57:24.939512 1 leaderelection.go:436] error retrieving resource lock kube-system/kube-controller-manager: Get "https://api-int.sno.openstack.lab:6443/apis/coordination.k8s.io/v1/namespaces/kube-system/leases/kube-controller-manager?timeout=6s": tls: failed to verify certificate: x509: certificate signed by unknown authority 2025-12-04T11:57:28.772338949Z E1204 11:57:28.772146 1 leaderelection.go:436] error retrieving resource lock kube-system/kube-controller-manager: Get "https://api-int.sno.openstack.lab:6443/apis/coordination.k8s.io/v1/namespaces/kube-system/leases/kube-controller-manager?timeout=6s": tls: failed to verify certificate: x509: certificate signed by unknown authority 2025-12-04T11:57:33.685241149Z E1204 11:57:33.685118 1 leaderelection.go:436] error retrieving resource lock kube-system/kube-controller-manager: Get "https://api-int.sno.openstack.lab:6443/apis/coordination.k8s.io/v1/namespaces/kube-system/leases/kube-controller-manager?timeout=6s": tls: failed to verify certificate: x509: certificate signed by unknown authority 2025-12-04T11:57:39.591698312Z E1204 11:57:39.591613 1 leaderelection.go:436] error retrieving resource lock kube-system/kube-controller-manager: Get "https://api-int.sno.openstack.lab:6443/apis/coordination.k8s.io/v1/namespaces/kube-system/leases/kube-controller-manager?timeout=6s": tls: failed to verify certificate: x509: certificate signed by unknown authority 2025-12-04T11:57:43.364680568Z E1204 11:57:43.364570 1 leaderelection.go:436] error retrieving resource lock kube-system/kube-controller-manager: Get "https://api-int.sno.openstack.lab:6443/apis/coordination.k8s.io/v1/namespaces/kube-system/leases/kube-controller-manager?timeout=6s": tls: failed to verify certificate: x509: certificate signed by unknown authority 2025-12-04T11:57:46.458652966Z E1204 11:57:46.458555 1 leaderelection.go:436] error retrieving resource lock kube-system/kube-controller-manager: Get "https://api-int.sno.openstack.lab:6443/apis/coordination.k8s.io/v1/namespaces/kube-system/leases/kube-controller-manager?timeout=6s": tls: failed to verify certificate: x509: certificate signed by unknown authority 2025-12-04T11:57:52.114864783Z E1204 11:57:52.114738 1 leaderelection.go:436] error retrieving resource lock kube-system/kube-controller-manager: Get "https://api-int.sno.openstack.lab:6443/apis/coordination.k8s.io/v1/namespaces/kube-system/leases/kube-controller-manager?timeout=6s": tls: failed to verify certificate: x509: certificate signed by unknown authority 2025-12-04T11:57:57.862451010Z E1204 11:57:57.862342 1 leaderelection.go:436] error retrieving resource lock kube-system/kube-controller-manager: Get "https://api-int.sno.openstack.lab:6443/apis/coordination.k8s.io/v1/namespaces/kube-system/leases/kube-controller-manager?timeout=6s": tls: failed to verify certificate: x509: certificate signed by unknown authority 2025-12-04T11:58:02.334563580Z E1204 11:58:02.334471 1 leaderelection.go:436] error retrieving resource lock kube-system/kube-controller-manager: Get "https://api-int.sno.openstack.lab:6443/apis/coordination.k8s.io/v1/namespaces/kube-system/leases/kube-controller-manager?timeout=6s": tls: failed to verify certificate: x509: certificate signed by unknown authority 2025-12-04T11:58:07.606933777Z E1204 11:58:07.606640 1 leaderelection.go:436] error retrieving resource lock kube-system/kube-controller-manager: Get "https://api-int.sno.openstack.lab:6443/apis/coordination.k8s.io/v1/namespaces/kube-system/leases/kube-controller-manager?timeout=6s": tls: failed to verify certificate: x509: certificate signed by unknown authority 2025-12-04T11:58:12.942915514Z E1204 11:58:12.942828 1 leaderelection.go:436] error retrieving resource lock kube-system/kube-controller-manager: Get "https://api-int.sno.openstack.lab:6443/apis/coordination.k8s.io/v1/namespaces/kube-system/leases/kube-controller-manager?timeout=6s": tls: failed to verify certificate: x509: certificate signed by unknown authority 2025-12-04T11:58:16.091954234Z E1204 11:58:16.091882 1 leaderelection.go:436] error retrieving resource lock kube-system/kube-controller-manager: Get "https://api-int.sno.openstack.lab:6443/apis/coordination.k8s.io/v1/namespaces/kube-system/leases/kube-controller-manager?timeout=6s": tls: failed to verify certificate: x509: certificate signed by unknown authority 2025-12-04T11:58:21.729890451Z E1204 11:58:21.729422 1 leaderelection.go:436] error retrieving resource lock kube-system/kube-controller-manager: Get "https://api-int.sno.openstack.lab:6443/apis/coordination.k8s.io/v1/namespaces/kube-system/leases/kube-controller-manager?timeout=6s": tls: failed to verify certificate: x509: certificate signed by unknown authority 2025-12-04T11:58:25.807214681Z E1204 11:58:25.807100 1 leaderelection.go:436] error retrieving resource lock kube-system/kube-controller-manager: Get "https://api-int.sno.openstack.lab:6443/apis/coordination.k8s.io/v1/namespaces/kube-system/leases/kube-controller-manager?timeout=6s": tls: failed to verify certificate: x509: certificate signed by unknown authority 2025-12-04T11:58:29.037607705Z E1204 11:58:29.037525 1 leaderelection.go:436] error retrieving resource lock kube-system/kube-controller-manager: Get "https://api-int.sno.openstack.lab:6443/apis/coordination.k8s.io/v1/namespaces/kube-system/leases/kube-controller-manager?timeout=6s": tls: failed to verify certificate: x509: certificate signed by unknown authority 2025-12-04T11:58:35.073438976Z E1204 11:58:35.073338 1 leaderelection.go:436] error retrieving resource lock kube-system/kube-controller-manager: Get "https://api-int.sno.openstack.lab:6443/apis/coordination.k8s.io/v1/namespaces/kube-system/leases/kube-controller-manager?timeout=6s": tls: failed to verify certificate: x509: certificate signed by unknown authority 2025-12-04T11:58:38.085197341Z E1204 11:58:38.085103 1 leaderelection.go:436] error retrieving resource lock kube-system/kube-controller-manager: Get "https://api-int.sno.openstack.lab:6443/apis/coordination.k8s.io/v1/namespaces/kube-system/leases/kube-controller-manager?timeout=6s": tls: failed to verify certificate: x509: certificate signed by unknown authority 2025-12-04T11:58:41.355415207Z E1204 11:58:41.355323 1 leaderelection.go:436] error retrieving resource lock kube-system/kube-controller-manager: Get "https://api-int.sno.openstack.lab:6443/apis/coordination.k8s.io/v1/namespaces/kube-system/leases/kube-controller-manager?timeout=6s": tls: failed to verify certificate: x509: certificate signed by unknown authority 2025-12-04T11:58:44.495840175Z E1204 11:58:44.495680 1 leaderelection.go:436] error retrieving resource lock kube-system/kube-controller-manager: Get "https://api-int.sno.openstack.lab:6443/apis/coordination.k8s.io/v1/namespaces/kube-system/leases/kube-controller-manager?timeout=6s": tls: failed to verify certificate: x509: certificate signed by unknown authority 2025-12-04T11:58:50.473753764Z E1204 11:58:50.473649 1 leaderelection.go:436] error retrieving resource lock kube-system/kube-controller-manager: Get "https://api-int.sno.openstack.lab:6443/apis/coordination.k8s.io/v1/namespaces/kube-system/leases/kube-controller-manager?timeout=6s": tls: failed to verify certificate: x509: certificate signed by unknown authority 2025-12-04T11:58:55.496372737Z E1204 11:58:55.496246 1 leaderelection.go:436] error retrieving resource lock kube-system/kube-controller-manager: Get "https://api-int.sno.openstack.lab:6443/apis/coordination.k8s.io/v1/namespaces/kube-system/leases/kube-controller-manager?timeout=6s": tls: failed to verify certificate: x509: certificate signed by unknown authority 2025-12-04T11:58:59.430762217Z E1204 11:58:59.430648 1 leaderelection.go:436] error retrieving resource lock kube-system/kube-controller-manager: Get "https://api-int.sno.openstack.lab:6443/apis/coordination.k8s.io/v1/namespaces/kube-system/leases/kube-controller-manager?timeout=6s": tls: failed to verify certificate: x509: certificate signed by unknown authority 2025-12-04T11:59:05.686042013Z E1204 11:59:05.685936 1 leaderelection.go:436] error retrieving resource lock kube-system/kube-controller-manager: Get "https://api-int.sno.openstack.lab:6443/apis/coordination.k8s.io/v1/namespaces/kube-system/leases/kube-controller-manager?timeout=6s": tls: failed to verify certificate: x509: certificate signed by unknown authority 2025-12-04T11:59:11.808201618Z E1204 11:59:11.808110 1 leaderelection.go:436] error retrieving resource lock kube-system/kube-controller-manager: Get "https://api-int.sno.openstack.lab:6443/apis/coordination.k8s.io/v1/namespaces/kube-system/leases/kube-controller-manager?timeout=6s": tls: failed to verify certificate: x509: certificate signed by unknown authority 2025-12-04T11:59:15.542883788Z E1204 11:59:15.542769 1 leaderelection.go:436] error retrieving resource lock kube-system/kube-controller-manager: Get "https://api-int.sno.openstack.lab:6443/apis/coordination.k8s.io/v1/namespaces/kube-system/leases/kube-controller-manager?timeout=6s": dial tcp 192.168.32.10:6443: connect: connection refused 2025-12-04T11:59:18.907464459Z E1204 11:59:18.907386 1 leaderelection.go:436] error retrieving resource lock kube-system/kube-controller-manager: Get "https://api-int.sno.openstack.lab:6443/apis/coordination.k8s.io/v1/namespaces/kube-system/leases/kube-controller-manager?timeout=6s": dial tcp 192.168.32.10:6443: connect: connection refused 2025-12-04T11:59:23.817612187Z E1204 11:59:23.817377 1 leaderelection.go:436] error retrieving resource lock kube-system/kube-controller-manager: Get "https://api-int.sno.openstack.lab:6443/apis/coordination.k8s.io/v1/namespaces/kube-system/leases/kube-controller-manager?timeout=6s": tls: failed to verify certificate: x509: certificate signed by unknown authority 2025-12-04T11:59:29.846392245Z E1204 11:59:29.846244 1 leaderelection.go:436] error retrieving resource lock kube-system/kube-controller-manager: Get "https://api-int.sno.openstack.lab:6443/apis/coordination.k8s.io/v1/namespaces/kube-system/leases/kube-controller-manager?timeout=6s": tls: failed to verify certificate: x509: certificate signed by unknown authority 2025-12-04T11:59:33.224970528Z E1204 11:59:33.224877 1 leaderelection.go:436] error retrieving resource lock kube-system/kube-controller-manager: Get "https://api-int.sno.openstack.lab:6443/apis/coordination.k8s.io/v1/namespaces/kube-system/leases/kube-controller-manager?timeout=6s": tls: failed to verify certificate: x509: certificate signed by unknown authority 2025-12-04T11:59:37.857538017Z E1204 11:59:37.857403 1 leaderelection.go:436] error retrieving resource lock kube-system/kube-controller-manager: Get "https://api-int.sno.openstack.lab:6443/apis/coordination.k8s.io/v1/namespaces/kube-system/leases/kube-controller-manager?timeout=6s": tls: failed to verify certificate: x509: certificate signed by unknown authority 2025-12-04T11:59:41.089863318Z E1204 11:59:41.089757 1 leaderelection.go:436] error retrieving resource lock kube-system/kube-controller-manager: Get "https://api-int.sno.openstack.lab:6443/apis/coordination.k8s.io/v1/namespaces/kube-system/leases/kube-controller-manager?timeout=6s": tls: failed to verify certificate: x509: certificate signed by unknown authority 2025-12-04T11:59:45.570073782Z E1204 11:59:45.569987 1 leaderelection.go:436] error retrieving resource lock kube-system/kube-controller-manager: Get "https://api-int.sno.openstack.lab:6443/apis/coordination.k8s.io/v1/namespaces/kube-system/leases/kube-controller-manager?timeout=6s": tls: failed to verify certificate: x509: certificate signed by unknown authority 2025-12-04T11:59:49.468662734Z E1204 11:59:49.468531 1 leaderelection.go:436] error retrieving resource lock kube-system/kube-controller-manager: Get "https://api-int.sno.openstack.lab:6443/apis/coordination.k8s.io/v1/namespaces/kube-system/leases/kube-controller-manager?timeout=6s": tls: failed to verify certificate: x509: certificate signed by unknown authority 2025-12-04T11:59:55.591675676Z E1204 11:59:55.590768 1 leaderelection.go:436] error retrieving resource lock kube-system/kube-controller-manager: Get "https://api-int.sno.openstack.lab:6443/apis/coordination.k8s.io/v1/namespaces/kube-system/leases/kube-controller-manager?timeout=6s": tls: failed to verify certificate: x509: certificate signed by unknown authority 2025-12-04T11:59:58.802900409Z E1204 11:59:58.802794 1 leaderelection.go:436] error retrieving resource lock kube-system/kube-controller-manager: Get "https://api-int.sno.openstack.lab:6443/apis/coordination.k8s.io/v1/namespaces/kube-system/leases/kube-controller-manager?timeout=6s": tls: failed to verify certificate: x509: certificate signed by unknown authority 2025-12-04T12:00:02.133834263Z E1204 12:00:02.133750 1 leaderelection.go:436] error retrieving resource lock kube-system/kube-controller-manager: Get "https://api-int.sno.openstack.lab:6443/apis/coordination.k8s.io/v1/namespaces/kube-system/leases/kube-controller-manager?timeout=6s": tls: failed to verify certificate: x509: certificate signed by unknown authority 2025-12-04T12:00:05.285067816Z E1204 12:00:05.284987 1 leaderelection.go:436] error retrieving resource lock kube-system/kube-controller-manager: Get "https://api-int.sno.openstack.lab:6443/apis/coordination.k8s.io/v1/namespaces/kube-system/leases/kube-controller-manager?timeout=6s": tls: failed to verify certificate: x509: certificate signed by unknown authority 2025-12-04T12:00:11.713285624Z E1204 12:00:11.713184 1 leaderelection.go:436] error retrieving resource lock kube-system/kube-controller-manager: Get "https://api-int.sno.openstack.lab:6443/apis/coordination.k8s.io/v1/namespaces/kube-system/leases/kube-controller-manager?timeout=6s": tls: failed to verify certificate: x509: certificate signed by unknown authority 2025-12-04T12:00:16.679295104Z E1204 12:00:16.679207 1 leaderelection.go:436] error retrieving resource lock kube-system/kube-controller-manager: Get "https://api-int.sno.openstack.lab:6443/apis/coordination.k8s.io/v1/namespaces/kube-system/leases/kube-controller-manager?timeout=6s": tls: failed to verify certificate: x509: certificate signed by unknown authority 2025-12-04T12:00:22.981372576Z E1204 12:00:22.981270 1 leaderelection.go:436] error retrieving resource lock kube-system/kube-controller-manager: Get "https://api-int.sno.openstack.lab:6443/apis/coordination.k8s.io/v1/namespaces/kube-system/leases/kube-controller-manager?timeout=6s": tls: failed to verify certificate: x509: certificate signed by unknown authority 2025-12-04T12:00:27.849520976Z E1204 12:00:27.849434 1 leaderelection.go:436] error retrieving resource lock kube-system/kube-controller-manager: Get "https://api-int.sno.openstack.lab:6443/apis/coordination.k8s.io/v1/namespaces/kube-system/leases/kube-controller-manager?timeout=6s": tls: failed to verify certificate: x509: certificate signed by unknown authority 2025-12-04T12:00:31.295424821Z E1204 12:00:31.295330 1 leaderelection.go:436] error retrieving resource lock kube-system/kube-controller-manager: Get "https://api-int.sno.openstack.lab:6443/apis/coordination.k8s.io/v1/namespaces/kube-system/leases/kube-controller-manager?timeout=6s": tls: failed to verify certificate: x509: certificate signed by unknown authority 2025-12-04T12:00:35.383178274Z E1204 12:00:35.383083 1 leaderelection.go:436] error retrieving resource lock kube-system/kube-controller-manager: Get "https://api-int.sno.openstack.lab:6443/apis/coordination.k8s.io/v1/namespaces/kube-system/leases/kube-controller-manager?timeout=6s": tls: failed to verify certificate: x509: certificate signed by unknown authority 2025-12-04T12:00:40.094745294Z E1204 12:00:40.094540 1 leaderelection.go:436] error retrieving resource lock kube-system/kube-controller-manager: Get "https://api-int.sno.openstack.lab:6443/apis/coordination.k8s.io/v1/namespaces/kube-system/leases/kube-controller-manager?timeout=6s": tls: failed to verify certificate: x509: certificate signed by unknown authority 2025-12-04T12:00:46.075573083Z E1204 12:00:46.075442 1 leaderelection.go:436] error retrieving resource lock kube-system/kube-controller-manager: Get "https://api-int.sno.openstack.lab:6443/apis/coordination.k8s.io/v1/namespaces/kube-system/leases/kube-controller-manager?timeout=6s": tls: failed to verify certificate: x509: certificate signed by unknown authority 2025-12-04T12:00:51.417190198Z E1204 12:00:51.417042 1 leaderelection.go:436] error retrieving resource lock kube-system/kube-controller-manager: Get "https://api-int.sno.openstack.lab:6443/apis/coordination.k8s.io/v1/namespaces/kube-system/leases/kube-controller-manager?timeout=6s": tls: failed to verify certificate: x509: certificate signed by unknown authority 2025-12-04T12:00:54.697579630Z E1204 12:00:54.697497 1 leaderelection.go:436] error retrieving resource lock kube-system/kube-controller-manager: Get "https://api-int.sno.openstack.lab:6443/apis/coordination.k8s.io/v1/namespaces/kube-system/leases/kube-controller-manager?timeout=6s": tls: failed to verify certificate: x509: certificate signed by unknown authority 2025-12-04T12:00:58.865485812Z E1204 12:00:58.865389 1 leaderelection.go:436] error retrieving resource lock kube-system/kube-controller-manager: Get "https://api-int.sno.openstack.lab:6443/apis/coordination.k8s.io/v1/namespaces/kube-system/leases/kube-controller-manager?timeout=6s": tls: failed to verify certificate: x509: certificate signed by unknown authority 2025-12-04T12:01:02.343728771Z E1204 12:01:02.343629 1 leaderelection.go:436] error retrieving resource lock kube-system/kube-controller-manager: Get "https://api-int.sno.openstack.lab:6443/apis/coordination.k8s.io/v1/namespaces/kube-system/leases/kube-controller-manager?timeout=6s": tls: failed to verify certificate: x509: certificate signed by unknown authority 2025-12-04T12:01:05.677223407Z E1204 12:01:05.677091 1 leaderelection.go:436] error retrieving resource lock kube-system/kube-controller-manager: Get "https://api-int.sno.openstack.lab:6443/apis/coordination.k8s.io/v1/namespaces/kube-system/leases/kube-controller-manager?timeout=6s": tls: failed to verify certificate: x509: certificate signed by unknown authority 2025-12-04T12:01:11.996561066Z E1204 12:01:11.996482 1 leaderelection.go:436] error retrieving resource lock kube-system/kube-controller-manager: Get "https://api-int.sno.openstack.lab:6443/apis/coordination.k8s.io/v1/namespaces/kube-system/leases/kube-controller-manager?timeout=6s": tls: failed to verify certificate: x509: certificate signed by unknown authority 2025-12-04T12:01:18.261063016Z E1204 12:01:18.260960 1 leaderelection.go:436] error retrieving resource lock kube-system/kube-controller-manager: Get "https://api-int.sno.openstack.lab:6443/apis/coordination.k8s.io/v1/namespaces/kube-system/leases/kube-controller-manager?timeout=6s": tls: failed to verify certificate: x509: certificate signed by unknown authority 2025-12-04T12:01:22.270968458Z E1204 12:01:22.270833 1 leaderelection.go:436] error retrieving resource lock kube-system/kube-controller-manager: Get "https://api-int.sno.openstack.lab:6443/apis/coordination.k8s.io/v1/namespaces/kube-system/leases/kube-controller-manager?timeout=6s": tls: failed to verify certificate: x509: certificate signed by unknown authority 2025-12-04T12:01:27.234492570Z E1204 12:01:27.234393 1 leaderelection.go:436] error retrieving resource lock kube-system/kube-controller-manager: Get "https://api-int.sno.openstack.lab:6443/apis/coordination.k8s.io/v1/namespaces/kube-system/leases/kube-controller-manager?timeout=6s": tls: failed to verify certificate: x509: certificate signed by unknown authority 2025-12-04T12:01:33.561939563Z E1204 12:01:33.561777 1 leaderelection.go:436] error retrieving resource lock kube-system/kube-controller-manager: Get "https://api-int.sno.openstack.lab:6443/apis/coordination.k8s.io/v1/namespaces/kube-system/leases/kube-controller-manager?timeout=6s": tls: failed to verify certificate: x509: certificate signed by unknown authority 2025-12-04T12:01:38.676449611Z E1204 12:01:38.675999 1 leaderelection.go:436] error retrieving resource lock kube-system/kube-controller-manager: Get "https://api-int.sno.openstack.lab:6443/apis/coordination.k8s.io/v1/namespaces/kube-system/leases/kube-controller-manager?timeout=6s": tls: failed to verify certificate: x509: certificate signed by unknown authority 2025-12-04T12:01:41.786783757Z E1204 12:01:41.786542 1 leaderelection.go:436] error retrieving resource lock kube-system/kube-controller-manager: Get "https://api-int.sno.openstack.lab:6443/apis/coordination.k8s.io/v1/namespaces/kube-system/leases/kube-controller-manager?timeout=6s": tls: failed to verify certificate: x509: certificate signed by unknown authority 2025-12-04T12:01:45.802742063Z E1204 12:01:45.802649 1 leaderelection.go:436] error retrieving resource lock kube-system/kube-controller-manager: Get "https://api-int.sno.openstack.lab:6443/apis/coordination.k8s.io/v1/namespaces/kube-system/leases/kube-controller-manager?timeout=6s": tls: failed to verify certificate: x509: certificate signed by unknown authority 2025-12-04T12:01:49.503900326Z E1204 12:01:49.503566 1 leaderelection.go:436] error retrieving resource lock kube-system/kube-controller-manager: Get "https://api-int.sno.openstack.lab:6443/apis/coordination.k8s.io/v1/namespaces/kube-system/leases/kube-controller-manager?timeout=6s": tls: failed to verify certificate: x509: certificate signed by unknown authority 2025-12-04T12:01:53.524617122Z E1204 12:01:53.524516 1 leaderelection.go:436] error retrieving resource lock kube-system/kube-controller-manager: Get "https://api-int.sno.openstack.lab:6443/apis/coordination.k8s.io/v1/namespaces/kube-system/leases/kube-controller-manager?timeout=6s": tls: failed to verify certificate: x509: certificate signed by unknown authority 2025-12-04T12:01:57.774325971Z E1204 12:01:57.773985 1 leaderelection.go:436] error retrieving resource lock kube-system/kube-controller-manager: Get "https://api-int.sno.openstack.lab:6443/apis/coordination.k8s.io/v1/namespaces/kube-system/leases/kube-controller-manager?timeout=6s": tls: failed to verify certificate: x509: certificate signed by unknown authority 2025-12-04T12:02:01.870004989Z E1204 12:02:01.869884 1 leaderelection.go:436] error retrieving resource lock kube-system/kube-controller-manager: Get "https://api-int.sno.openstack.lab:6443/apis/coordination.k8s.io/v1/namespaces/kube-system/leases/kube-controller-manager?timeout=6s": tls: failed to verify certificate: x509: certificate signed by unknown authority kube-controller-manager-cert-syncer_previous.log0000640000000000000000000012474115114274105040460 0ustar0000000000000000config/pod/openshift-kube-controller-manager/logs/kube-controller-manager-master-02025-12-04T11:45:45.586786153Z I1204 11:45:45.586331 1 observer_polling.go:159] Starting file observer 2025-12-04T11:45:45.589447119Z I1204 11:45:45.587301 1 base_controller.go:76] Waiting for caches to sync for CertSyncController 2025-12-04T11:45:45.592534430Z W1204 11:45:45.592452 1 reflector.go:561] k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243: failed to list *v1.ConfigMap: Get "https://localhost:6443/api/v1/namespaces/openshift-kube-controller-manager/configmaps?limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate is valid for kubernetes, kubernetes.default, kubernetes.default.svc, kubernetes.default.svc.cluster.local, openshift, openshift.default, openshift.default.svc, openshift.default.svc.cluster.local, not localhost-recovery 2025-12-04T11:45:45.592683165Z E1204 11:45:45.592647 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243: Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: Get \"https://localhost:6443/api/v1/namespaces/openshift-kube-controller-manager/configmaps?limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate is valid for kubernetes, kubernetes.default, kubernetes.default.svc, kubernetes.default.svc.cluster.local, openshift, openshift.default, openshift.default.svc, openshift.default.svc.cluster.local, not localhost-recovery" 2025-12-04T11:45:45.593076958Z W1204 11:45:45.592985 1 reflector.go:561] k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243: failed to list *v1.Secret: Get "https://localhost:6443/api/v1/namespaces/openshift-kube-controller-manager/secrets?limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate is valid for kubernetes, kubernetes.default, kubernetes.default.svc, kubernetes.default.svc.cluster.local, openshift, openshift.default, openshift.default.svc, openshift.default.svc.cluster.local, not localhost-recovery 2025-12-04T11:45:45.593129430Z E1204 11:45:45.593103 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243: Failed to watch *v1.Secret: failed to list *v1.Secret: Get \"https://localhost:6443/api/v1/namespaces/openshift-kube-controller-manager/secrets?limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate is valid for kubernetes, kubernetes.default, kubernetes.default.svc, kubernetes.default.svc.cluster.local, openshift, openshift.default, openshift.default.svc, openshift.default.svc.cluster.local, not localhost-recovery" 2025-12-04T11:45:46.473483893Z W1204 11:45:46.473338 1 reflector.go:561] k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243: failed to list *v1.ConfigMap: Get "https://localhost:6443/api/v1/namespaces/openshift-kube-controller-manager/configmaps?limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate is valid for kubernetes, kubernetes.default, kubernetes.default.svc, kubernetes.default.svc.cluster.local, openshift, openshift.default, openshift.default.svc, openshift.default.svc.cluster.local, not localhost-recovery 2025-12-04T11:45:46.473676689Z E1204 11:45:46.473640 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243: Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: Get \"https://localhost:6443/api/v1/namespaces/openshift-kube-controller-manager/configmaps?limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate is valid for kubernetes, kubernetes.default, kubernetes.default.svc, kubernetes.default.svc.cluster.local, openshift, openshift.default, openshift.default.svc, openshift.default.svc.cluster.local, not localhost-recovery" 2025-12-04T11:45:46.923453654Z W1204 11:45:46.923278 1 reflector.go:561] k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243: failed to list *v1.Secret: Get "https://localhost:6443/api/v1/namespaces/openshift-kube-controller-manager/secrets?limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate is valid for kubernetes, kubernetes.default, kubernetes.default.svc, kubernetes.default.svc.cluster.local, openshift, openshift.default, openshift.default.svc, openshift.default.svc.cluster.local, not localhost-recovery 2025-12-04T11:45:46.923453654Z E1204 11:45:46.923420 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243: Failed to watch *v1.Secret: failed to list *v1.Secret: Get \"https://localhost:6443/api/v1/namespaces/openshift-kube-controller-manager/secrets?limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate is valid for kubernetes, kubernetes.default, kubernetes.default.svc, kubernetes.default.svc.cluster.local, openshift, openshift.default, openshift.default.svc, openshift.default.svc.cluster.local, not localhost-recovery" 2025-12-04T11:45:48.569419865Z W1204 11:45:48.569296 1 reflector.go:561] k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243: failed to list *v1.ConfigMap: Get "https://localhost:6443/api/v1/namespaces/openshift-kube-controller-manager/configmaps?limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate is valid for kubernetes, kubernetes.default, kubernetes.default.svc, kubernetes.default.svc.cluster.local, openshift, openshift.default, openshift.default.svc, openshift.default.svc.cluster.local, not localhost-recovery 2025-12-04T11:45:48.569470857Z E1204 11:45:48.569406 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243: Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: Get \"https://localhost:6443/api/v1/namespaces/openshift-kube-controller-manager/configmaps?limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate is valid for kubernetes, kubernetes.default, kubernetes.default.svc, kubernetes.default.svc.cluster.local, openshift, openshift.default, openshift.default.svc, openshift.default.svc.cluster.local, not localhost-recovery" 2025-12-04T11:45:49.798979980Z W1204 11:45:49.798884 1 reflector.go:561] k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243: failed to list *v1.Secret: Get "https://localhost:6443/api/v1/namespaces/openshift-kube-controller-manager/secrets?limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate is valid for kubernetes, kubernetes.default, kubernetes.default.svc, kubernetes.default.svc.cluster.local, openshift, openshift.default, openshift.default.svc, openshift.default.svc.cluster.local, not localhost-recovery 2025-12-04T11:45:49.798979980Z E1204 11:45:49.798951 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243: Failed to watch *v1.Secret: failed to list *v1.Secret: Get \"https://localhost:6443/api/v1/namespaces/openshift-kube-controller-manager/secrets?limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate is valid for kubernetes, kubernetes.default, kubernetes.default.svc, kubernetes.default.svc.cluster.local, openshift, openshift.default, openshift.default.svc, openshift.default.svc.cluster.local, not localhost-recovery" 2025-12-04T11:45:52.140744409Z W1204 11:45:52.140605 1 reflector.go:561] k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243: failed to list *v1.ConfigMap: Get "https://localhost:6443/api/v1/namespaces/openshift-kube-controller-manager/configmaps?limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate is valid for kubernetes, kubernetes.default, kubernetes.default.svc, kubernetes.default.svc.cluster.local, openshift, openshift.default, openshift.default.svc, openshift.default.svc.cluster.local, not localhost-recovery 2025-12-04T11:45:52.140744409Z E1204 11:45:52.140697 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243: Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: Get \"https://localhost:6443/api/v1/namespaces/openshift-kube-controller-manager/configmaps?limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate is valid for kubernetes, kubernetes.default, kubernetes.default.svc, kubernetes.default.svc.cluster.local, openshift, openshift.default, openshift.default.svc, openshift.default.svc.cluster.local, not localhost-recovery" 2025-12-04T11:45:54.620155642Z W1204 11:45:54.619990 1 reflector.go:561] k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243: failed to list *v1.Secret: Get "https://localhost:6443/api/v1/namespaces/openshift-kube-controller-manager/secrets?limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate is valid for kubernetes, kubernetes.default, kubernetes.default.svc, kubernetes.default.svc.cluster.local, openshift, openshift.default, openshift.default.svc, openshift.default.svc.cluster.local, not localhost-recovery 2025-12-04T11:45:54.620155642Z E1204 11:45:54.620074 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243: Failed to watch *v1.Secret: failed to list *v1.Secret: Get \"https://localhost:6443/api/v1/namespaces/openshift-kube-controller-manager/secrets?limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate is valid for kubernetes, kubernetes.default, kubernetes.default.svc, kubernetes.default.svc.cluster.local, openshift, openshift.default, openshift.default.svc, openshift.default.svc.cluster.local, not localhost-recovery" 2025-12-04T11:46:01.836338770Z W1204 11:46:01.836237 1 reflector.go:561] k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243: failed to list *v1.ConfigMap: Get "https://localhost:6443/api/v1/namespaces/openshift-kube-controller-manager/configmaps?limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate is valid for kubernetes, kubernetes.default, kubernetes.default.svc, kubernetes.default.svc.cluster.local, openshift, openshift.default, openshift.default.svc, openshift.default.svc.cluster.local, not localhost-recovery 2025-12-04T11:46:01.836338770Z E1204 11:46:01.836317 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243: Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: Get \"https://localhost:6443/api/v1/namespaces/openshift-kube-controller-manager/configmaps?limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate is valid for kubernetes, kubernetes.default, kubernetes.default.svc, kubernetes.default.svc.cluster.local, openshift, openshift.default, openshift.default.svc, openshift.default.svc.cluster.local, not localhost-recovery" 2025-12-04T11:46:04.618732974Z W1204 11:46:04.618602 1 reflector.go:561] k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243: failed to list *v1.Secret: Get "https://localhost:6443/api/v1/namespaces/openshift-kube-controller-manager/secrets?limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate is valid for kubernetes, kubernetes.default, kubernetes.default.svc, kubernetes.default.svc.cluster.local, openshift, openshift.default, openshift.default.svc, openshift.default.svc.cluster.local, not localhost-recovery 2025-12-04T11:46:04.618732974Z E1204 11:46:04.618677 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243: Failed to watch *v1.Secret: failed to list *v1.Secret: Get \"https://localhost:6443/api/v1/namespaces/openshift-kube-controller-manager/secrets?limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate is valid for kubernetes, kubernetes.default, kubernetes.default.svc, kubernetes.default.svc.cluster.local, openshift, openshift.default, openshift.default.svc, openshift.default.svc.cluster.local, not localhost-recovery" 2025-12-04T11:46:17.106299071Z W1204 11:46:17.106191 1 reflector.go:561] k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243: failed to list *v1.ConfigMap: Get "https://localhost:6443/api/v1/namespaces/openshift-kube-controller-manager/configmaps?limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate is valid for kubernetes, kubernetes.default, kubernetes.default.svc, kubernetes.default.svc.cluster.local, openshift, openshift.default, openshift.default.svc, openshift.default.svc.cluster.local, not localhost-recovery 2025-12-04T11:46:17.106299071Z E1204 11:46:17.106251 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243: Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: Get \"https://localhost:6443/api/v1/namespaces/openshift-kube-controller-manager/configmaps?limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate is valid for kubernetes, kubernetes.default, kubernetes.default.svc, kubernetes.default.svc.cluster.local, openshift, openshift.default, openshift.default.svc, openshift.default.svc.cluster.local, not localhost-recovery" 2025-12-04T11:46:22.768056982Z W1204 11:46:22.767939 1 reflector.go:561] k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243: failed to list *v1.Secret: Get "https://localhost:6443/api/v1/namespaces/openshift-kube-controller-manager/secrets?limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate is valid for kubernetes, kubernetes.default, kubernetes.default.svc, kubernetes.default.svc.cluster.local, openshift, openshift.default, openshift.default.svc, openshift.default.svc.cluster.local, not localhost-recovery 2025-12-04T11:46:22.768056982Z E1204 11:46:22.768005 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243: Failed to watch *v1.Secret: failed to list *v1.Secret: Get \"https://localhost:6443/api/v1/namespaces/openshift-kube-controller-manager/secrets?limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate is valid for kubernetes, kubernetes.default, kubernetes.default.svc, kubernetes.default.svc.cluster.local, openshift, openshift.default, openshift.default.svc, openshift.default.svc.cluster.local, not localhost-recovery" 2025-12-04T11:47:02.832223236Z W1204 11:47:02.832058 1 reflector.go:561] k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243: failed to list *v1.ConfigMap: Get "https://localhost:6443/api/v1/namespaces/openshift-kube-controller-manager/configmaps?limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate is valid for kubernetes, kubernetes.default, kubernetes.default.svc, kubernetes.default.svc.cluster.local, openshift, openshift.default, openshift.default.svc, openshift.default.svc.cluster.local, not localhost-recovery 2025-12-04T11:47:02.832223236Z E1204 11:47:02.832172 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243: Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: Get \"https://localhost:6443/api/v1/namespaces/openshift-kube-controller-manager/configmaps?limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate is valid for kubernetes, kubernetes.default, kubernetes.default.svc, kubernetes.default.svc.cluster.local, openshift, openshift.default, openshift.default.svc, openshift.default.svc.cluster.local, not localhost-recovery" 2025-12-04T11:47:03.330523492Z W1204 11:47:03.330382 1 reflector.go:561] k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243: failed to list *v1.Secret: Get "https://localhost:6443/api/v1/namespaces/openshift-kube-controller-manager/secrets?limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate is valid for kubernetes, kubernetes.default, kubernetes.default.svc, kubernetes.default.svc.cluster.local, openshift, openshift.default, openshift.default.svc, openshift.default.svc.cluster.local, not localhost-recovery 2025-12-04T11:47:03.330523492Z E1204 11:47:03.330485 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243: Failed to watch *v1.Secret: failed to list *v1.Secret: Get \"https://localhost:6443/api/v1/namespaces/openshift-kube-controller-manager/secrets?limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate is valid for kubernetes, kubernetes.default, kubernetes.default.svc, kubernetes.default.svc.cluster.local, openshift, openshift.default, openshift.default.svc, openshift.default.svc.cluster.local, not localhost-recovery" 2025-12-04T11:47:33.902271934Z W1204 11:47:33.902185 1 reflector.go:561] k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243: failed to list *v1.ConfigMap: Get "https://localhost:6443/api/v1/namespaces/openshift-kube-controller-manager/configmaps?limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate is valid for kubernetes, kubernetes.default, kubernetes.default.svc, kubernetes.default.svc.cluster.local, openshift, openshift.default, openshift.default.svc, openshift.default.svc.cluster.local, not localhost-recovery 2025-12-04T11:47:33.902422109Z E1204 11:47:33.902406 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243: Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: Get \"https://localhost:6443/api/v1/namespaces/openshift-kube-controller-manager/configmaps?limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate is valid for kubernetes, kubernetes.default, kubernetes.default.svc, kubernetes.default.svc.cluster.local, openshift, openshift.default, openshift.default.svc, openshift.default.svc.cluster.local, not localhost-recovery" 2025-12-04T11:48:00.252588174Z W1204 11:48:00.252463 1 reflector.go:561] k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243: failed to list *v1.Secret: Get "https://localhost:6443/api/v1/namespaces/openshift-kube-controller-manager/secrets?limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate is valid for kubernetes, kubernetes.default, kubernetes.default.svc, kubernetes.default.svc.cluster.local, openshift, openshift.default, openshift.default.svc, openshift.default.svc.cluster.local, not localhost-recovery 2025-12-04T11:48:00.252768410Z E1204 11:48:00.252747 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243: Failed to watch *v1.Secret: failed to list *v1.Secret: Get \"https://localhost:6443/api/v1/namespaces/openshift-kube-controller-manager/secrets?limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate is valid for kubernetes, kubernetes.default, kubernetes.default.svc, kubernetes.default.svc.cluster.local, openshift, openshift.default, openshift.default.svc, openshift.default.svc.cluster.local, not localhost-recovery" 2025-12-04T11:48:26.296003611Z W1204 11:48:26.295700 1 reflector.go:561] k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243: failed to list *v1.ConfigMap: Get "https://localhost:6443/api/v1/namespaces/openshift-kube-controller-manager/configmaps?limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate is valid for kubernetes, kubernetes.default, kubernetes.default.svc, kubernetes.default.svc.cluster.local, openshift, openshift.default, openshift.default.svc, openshift.default.svc.cluster.local, not localhost-recovery 2025-12-04T11:48:26.296003611Z E1204 11:48:26.295828 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243: Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: Get \"https://localhost:6443/api/v1/namespaces/openshift-kube-controller-manager/configmaps?limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate is valid for kubernetes, kubernetes.default, kubernetes.default.svc, kubernetes.default.svc.cluster.local, openshift, openshift.default, openshift.default.svc, openshift.default.svc.cluster.local, not localhost-recovery" 2025-12-04T11:48:42.087147373Z W1204 11:48:42.087045 1 reflector.go:561] k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243: failed to list *v1.Secret: Get "https://localhost:6443/api/v1/namespaces/openshift-kube-controller-manager/secrets?limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate is valid for kubernetes, kubernetes.default, kubernetes.default.svc, kubernetes.default.svc.cluster.local, openshift, openshift.default, openshift.default.svc, openshift.default.svc.cluster.local, not localhost-recovery 2025-12-04T11:48:42.087212196Z E1204 11:48:42.087151 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243: Failed to watch *v1.Secret: failed to list *v1.Secret: Get \"https://localhost:6443/api/v1/namespaces/openshift-kube-controller-manager/secrets?limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate is valid for kubernetes, kubernetes.default, kubernetes.default.svc, kubernetes.default.svc.cluster.local, openshift, openshift.default, openshift.default.svc, openshift.default.svc.cluster.local, not localhost-recovery" 2025-12-04T11:49:20.691774629Z W1204 11:49:20.691618 1 reflector.go:561] k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243: failed to list *v1.ConfigMap: Get "https://localhost:6443/api/v1/namespaces/openshift-kube-controller-manager/configmaps?limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate is valid for kubernetes, kubernetes.default, kubernetes.default.svc, kubernetes.default.svc.cluster.local, openshift, openshift.default, openshift.default.svc, openshift.default.svc.cluster.local, not localhost-recovery 2025-12-04T11:49:20.691774629Z E1204 11:49:20.691743 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243: Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: Get \"https://localhost:6443/api/v1/namespaces/openshift-kube-controller-manager/configmaps?limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate is valid for kubernetes, kubernetes.default, kubernetes.default.svc, kubernetes.default.svc.cluster.local, openshift, openshift.default, openshift.default.svc, openshift.default.svc.cluster.local, not localhost-recovery" 2025-12-04T11:49:25.727327988Z W1204 11:49:25.727206 1 reflector.go:561] k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243: failed to list *v1.Secret: Get "https://localhost:6443/api/v1/namespaces/openshift-kube-controller-manager/secrets?limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate is valid for kubernetes, kubernetes.default, kubernetes.default.svc, kubernetes.default.svc.cluster.local, openshift, openshift.default, openshift.default.svc, openshift.default.svc.cluster.local, not localhost-recovery 2025-12-04T11:49:25.727327988Z E1204 11:49:25.727293 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243: Failed to watch *v1.Secret: failed to list *v1.Secret: Get \"https://localhost:6443/api/v1/namespaces/openshift-kube-controller-manager/secrets?limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate is valid for kubernetes, kubernetes.default, kubernetes.default.svc, kubernetes.default.svc.cluster.local, openshift, openshift.default, openshift.default.svc, openshift.default.svc.cluster.local, not localhost-recovery" 2025-12-04T11:49:53.144994149Z W1204 11:49:53.144884 1 reflector.go:561] k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243: failed to list *v1.ConfigMap: Get "https://localhost:6443/api/v1/namespaces/openshift-kube-controller-manager/configmaps?limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate is valid for kubernetes, kubernetes.default, kubernetes.default.svc, kubernetes.default.svc.cluster.local, openshift, openshift.default, openshift.default.svc, openshift.default.svc.cluster.local, not localhost-recovery 2025-12-04T11:49:53.145076672Z E1204 11:49:53.144988 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243: Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: Get \"https://localhost:6443/api/v1/namespaces/openshift-kube-controller-manager/configmaps?limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate is valid for kubernetes, kubernetes.default, kubernetes.default.svc, kubernetes.default.svc.cluster.local, openshift, openshift.default, openshift.default.svc, openshift.default.svc.cluster.local, not localhost-recovery" 2025-12-04T11:50:12.084864183Z W1204 11:50:12.083799 1 reflector.go:561] k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243: failed to list *v1.Secret: Get "https://localhost:6443/api/v1/namespaces/openshift-kube-controller-manager/secrets?limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate is valid for kubernetes, kubernetes.default, kubernetes.default.svc, kubernetes.default.svc.cluster.local, openshift, openshift.default, openshift.default.svc, openshift.default.svc.cluster.local, not localhost-recovery 2025-12-04T11:50:12.084864183Z E1204 11:50:12.084017 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243: Failed to watch *v1.Secret: failed to list *v1.Secret: Get \"https://localhost:6443/api/v1/namespaces/openshift-kube-controller-manager/secrets?limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate is valid for kubernetes, kubernetes.default, kubernetes.default.svc, kubernetes.default.svc.cluster.local, openshift, openshift.default, openshift.default.svc, openshift.default.svc.cluster.local, not localhost-recovery" 2025-12-04T11:50:30.724932119Z W1204 11:50:30.724765 1 reflector.go:561] k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243: failed to list *v1.ConfigMap: Get "https://localhost:6443/api/v1/namespaces/openshift-kube-controller-manager/configmaps?limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate is valid for kubernetes, kubernetes.default, kubernetes.default.svc, kubernetes.default.svc.cluster.local, openshift, openshift.default, openshift.default.svc, openshift.default.svc.cluster.local, not localhost-recovery 2025-12-04T11:50:30.724932119Z E1204 11:50:30.724892 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243: Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: Get \"https://localhost:6443/api/v1/namespaces/openshift-kube-controller-manager/configmaps?limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate is valid for kubernetes, kubernetes.default, kubernetes.default.svc, kubernetes.default.svc.cluster.local, openshift, openshift.default, openshift.default.svc, openshift.default.svc.cluster.local, not localhost-recovery" 2025-12-04T11:50:42.611095794Z W1204 11:50:42.610976 1 reflector.go:561] k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243: failed to list *v1.Secret: Get "https://localhost:6443/api/v1/namespaces/openshift-kube-controller-manager/secrets?limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate is valid for kubernetes, kubernetes.default, kubernetes.default.svc, kubernetes.default.svc.cluster.local, openshift, openshift.default, openshift.default.svc, openshift.default.svc.cluster.local, not localhost-recovery 2025-12-04T11:50:42.611095794Z E1204 11:50:42.611052 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243: Failed to watch *v1.Secret: failed to list *v1.Secret: Get \"https://localhost:6443/api/v1/namespaces/openshift-kube-controller-manager/secrets?limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate is valid for kubernetes, kubernetes.default, kubernetes.default.svc, kubernetes.default.svc.cluster.local, openshift, openshift.default, openshift.default.svc, openshift.default.svc.cluster.local, not localhost-recovery" 2025-12-04T11:51:10.390575203Z W1204 11:51:10.390448 1 reflector.go:561] k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243: failed to list *v1.ConfigMap: Get "https://localhost:6443/api/v1/namespaces/openshift-kube-controller-manager/configmaps?limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate is valid for kubernetes, kubernetes.default, kubernetes.default.svc, kubernetes.default.svc.cluster.local, openshift, openshift.default, openshift.default.svc, openshift.default.svc.cluster.local, not localhost-recovery 2025-12-04T11:51:10.390575203Z E1204 11:51:10.390558 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243: Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: Get \"https://localhost:6443/api/v1/namespaces/openshift-kube-controller-manager/configmaps?limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate is valid for kubernetes, kubernetes.default, kubernetes.default.svc, kubernetes.default.svc.cluster.local, openshift, openshift.default, openshift.default.svc, openshift.default.svc.cluster.local, not localhost-recovery" 2025-12-04T11:51:13.400833834Z W1204 11:51:13.400731 1 reflector.go:561] k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243: failed to list *v1.Secret: Get "https://localhost:6443/api/v1/namespaces/openshift-kube-controller-manager/secrets?limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate is valid for kubernetes, kubernetes.default, kubernetes.default.svc, kubernetes.default.svc.cluster.local, openshift, openshift.default, openshift.default.svc, openshift.default.svc.cluster.local, not localhost-recovery 2025-12-04T11:51:13.400833834Z E1204 11:51:13.400825 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243: Failed to watch *v1.Secret: failed to list *v1.Secret: Get \"https://localhost:6443/api/v1/namespaces/openshift-kube-controller-manager/secrets?limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate is valid for kubernetes, kubernetes.default, kubernetes.default.svc, kubernetes.default.svc.cluster.local, openshift, openshift.default, openshift.default.svc, openshift.default.svc.cluster.local, not localhost-recovery" 2025-12-04T11:51:44.132523725Z W1204 11:51:44.132408 1 reflector.go:561] k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243: failed to list *v1.ConfigMap: Get "https://localhost:6443/api/v1/namespaces/openshift-kube-controller-manager/configmaps?limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate is valid for kubernetes, kubernetes.default, kubernetes.default.svc, kubernetes.default.svc.cluster.local, openshift, openshift.default, openshift.default.svc, openshift.default.svc.cluster.local, not localhost-recovery 2025-12-04T11:51:44.132614798Z E1204 11:51:44.132535 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243: Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: Get \"https://localhost:6443/api/v1/namespaces/openshift-kube-controller-manager/configmaps?limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate is valid for kubernetes, kubernetes.default, kubernetes.default.svc, kubernetes.default.svc.cluster.local, openshift, openshift.default, openshift.default.svc, openshift.default.svc.cluster.local, not localhost-recovery" 2025-12-04T11:52:00.677928615Z W1204 11:52:00.677822 1 reflector.go:561] k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243: failed to list *v1.Secret: Get "https://localhost:6443/api/v1/namespaces/openshift-kube-controller-manager/secrets?limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate is valid for kubernetes, kubernetes.default, kubernetes.default.svc, kubernetes.default.svc.cluster.local, openshift, openshift.default, openshift.default.svc, openshift.default.svc.cluster.local, not localhost-recovery 2025-12-04T11:52:00.677928615Z E1204 11:52:00.677899 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243: Failed to watch *v1.Secret: failed to list *v1.Secret: Get \"https://localhost:6443/api/v1/namespaces/openshift-kube-controller-manager/secrets?limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate is valid for kubernetes, kubernetes.default, kubernetes.default.svc, kubernetes.default.svc.cluster.local, openshift, openshift.default, openshift.default.svc, openshift.default.svc.cluster.local, not localhost-recovery" 2025-12-04T11:52:19.721460172Z W1204 11:52:19.721370 1 reflector.go:561] k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243: failed to list *v1.ConfigMap: Get "https://localhost:6443/api/v1/namespaces/openshift-kube-controller-manager/configmaps?limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate is valid for kubernetes, kubernetes.default, kubernetes.default.svc, kubernetes.default.svc.cluster.local, openshift, openshift.default, openshift.default.svc, openshift.default.svc.cluster.local, not localhost-recovery 2025-12-04T11:52:19.721460172Z E1204 11:52:19.721433 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243: Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: Get \"https://localhost:6443/api/v1/namespaces/openshift-kube-controller-manager/configmaps?limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate is valid for kubernetes, kubernetes.default, kubernetes.default.svc, kubernetes.default.svc.cluster.local, openshift, openshift.default, openshift.default.svc, openshift.default.svc.cluster.local, not localhost-recovery" 2025-12-04T11:52:35.836935767Z W1204 11:52:35.836755 1 reflector.go:561] k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243: failed to list *v1.Secret: Get "https://localhost:6443/api/v1/namespaces/openshift-kube-controller-manager/secrets?limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate is valid for kubernetes, kubernetes.default, kubernetes.default.svc, kubernetes.default.svc.cluster.local, openshift, openshift.default, openshift.default.svc, openshift.default.svc.cluster.local, not localhost-recovery 2025-12-04T11:52:35.836994109Z E1204 11:52:35.836932 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243: Failed to watch *v1.Secret: failed to list *v1.Secret: Get \"https://localhost:6443/api/v1/namespaces/openshift-kube-controller-manager/secrets?limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate is valid for kubernetes, kubernetes.default, kubernetes.default.svc, kubernetes.default.svc.cluster.local, openshift, openshift.default, openshift.default.svc, openshift.default.svc.cluster.local, not localhost-recovery" 2025-12-04T11:53:05.160531275Z W1204 11:53:05.160380 1 reflector.go:561] k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243: failed to list *v1.ConfigMap: Get "https://localhost:6443/api/v1/namespaces/openshift-kube-controller-manager/configmaps?limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate is valid for kubernetes, kubernetes.default, kubernetes.default.svc, kubernetes.default.svc.cluster.local, openshift, openshift.default, openshift.default.svc, openshift.default.svc.cluster.local, not localhost-recovery 2025-12-04T11:53:05.160617018Z E1204 11:53:05.160529 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243: Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: Get \"https://localhost:6443/api/v1/namespaces/openshift-kube-controller-manager/configmaps?limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate is valid for kubernetes, kubernetes.default, kubernetes.default.svc, kubernetes.default.svc.cluster.local, openshift, openshift.default, openshift.default.svc, openshift.default.svc.cluster.local, not localhost-recovery" 2025-12-04T11:53:06.556682506Z W1204 11:53:06.556553 1 reflector.go:561] k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243: failed to list *v1.Secret: Get "https://localhost:6443/api/v1/namespaces/openshift-kube-controller-manager/secrets?limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate is valid for kubernetes, kubernetes.default, kubernetes.default.svc, kubernetes.default.svc.cluster.local, openshift, openshift.default, openshift.default.svc, openshift.default.svc.cluster.local, not localhost-recovery 2025-12-04T11:53:06.556682506Z E1204 11:53:06.556629 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243: Failed to watch *v1.Secret: failed to list *v1.Secret: Get \"https://localhost:6443/api/v1/namespaces/openshift-kube-controller-manager/secrets?limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate is valid for kubernetes, kubernetes.default, kubernetes.default.svc, kubernetes.default.svc.cluster.local, openshift, openshift.default, openshift.default.svc, openshift.default.svc.cluster.local, not localhost-recovery" 2025-12-04T11:54:03.744078329Z W1204 11:54:03.743960 1 reflector.go:561] k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243: failed to list *v1.ConfigMap: Get "https://localhost:6443/api/v1/namespaces/openshift-kube-controller-manager/configmaps?limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate is valid for kubernetes, kubernetes.default, kubernetes.default.svc, kubernetes.default.svc.cluster.local, openshift, openshift.default, openshift.default.svc, openshift.default.svc.cluster.local, not localhost-recovery 2025-12-04T11:54:03.744078329Z E1204 11:54:03.744064 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243: Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: Get \"https://localhost:6443/api/v1/namespaces/openshift-kube-controller-manager/configmaps?limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate is valid for kubernetes, kubernetes.default, kubernetes.default.svc, kubernetes.default.svc.cluster.local, openshift, openshift.default, openshift.default.svc, openshift.default.svc.cluster.local, not localhost-recovery" 2025-12-04T11:54:06.009337183Z W1204 11:54:06.009199 1 reflector.go:561] k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243: failed to list *v1.Secret: Get "https://localhost:6443/api/v1/namespaces/openshift-kube-controller-manager/secrets?limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate is valid for kubernetes, kubernetes.default, kubernetes.default.svc, kubernetes.default.svc.cluster.local, openshift, openshift.default, openshift.default.svc, openshift.default.svc.cluster.local, not localhost-recovery 2025-12-04T11:54:06.009337183Z E1204 11:54:06.009301 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243: Failed to watch *v1.Secret: failed to list *v1.Secret: Get \"https://localhost:6443/api/v1/namespaces/openshift-kube-controller-manager/secrets?limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate is valid for kubernetes, kubernetes.default, kubernetes.default.svc, kubernetes.default.svc.cluster.local, openshift, openshift.default, openshift.default.svc, openshift.default.svc.cluster.local, not localhost-recovery" 2025-12-04T11:54:59.810636954Z W1204 11:54:59.810467 1 reflector.go:561] k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243: failed to list *v1.Secret: Get "https://localhost:6443/api/v1/namespaces/openshift-kube-controller-manager/secrets?limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate is valid for kubernetes, kubernetes.default, kubernetes.default.svc, kubernetes.default.svc.cluster.local, openshift, openshift.default, openshift.default.svc, openshift.default.svc.cluster.local, not localhost-recovery 2025-12-04T11:54:59.810894822Z E1204 11:54:59.810874 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243: Failed to watch *v1.Secret: failed to list *v1.Secret: Get \"https://localhost:6443/api/v1/namespaces/openshift-kube-controller-manager/secrets?limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate is valid for kubernetes, kubernetes.default, kubernetes.default.svc, kubernetes.default.svc.cluster.local, openshift, openshift.default, openshift.default.svc, openshift.default.svc.cluster.local, not localhost-recovery" 2025-12-04T11:55:00.404443377Z W1204 11:55:00.404291 1 reflector.go:561] k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243: failed to list *v1.ConfigMap: Get "https://localhost:6443/api/v1/namespaces/openshift-kube-controller-manager/configmaps?limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate is valid for kubernetes, kubernetes.default, kubernetes.default.svc, kubernetes.default.svc.cluster.local, openshift, openshift.default, openshift.default.svc, openshift.default.svc.cluster.local, not localhost-recovery 2025-12-04T11:55:00.404443377Z E1204 11:55:00.404388 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243: Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: Get \"https://localhost:6443/api/v1/namespaces/openshift-kube-controller-manager/configmaps?limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate is valid for kubernetes, kubernetes.default, kubernetes.default.svc, kubernetes.default.svc.cluster.local, openshift, openshift.default, openshift.default.svc, openshift.default.svc.cluster.local, not localhost-recovery" 2025-12-04T11:55:40.529992107Z W1204 11:55:40.529780 1 reflector.go:561] k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243: failed to list *v1.Secret: Get "https://localhost:6443/api/v1/namespaces/openshift-kube-controller-manager/secrets?limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate is valid for kubernetes, kubernetes.default, kubernetes.default.svc, kubernetes.default.svc.cluster.local, openshift, openshift.default, openshift.default.svc, openshift.default.svc.cluster.local, not localhost-recovery 2025-12-04T11:55:40.530111331Z E1204 11:55:40.529981 1 reflector.go:158] "Unhandled Error" err="k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243: Failed to watch *v1.Secret: failed to list *v1.Secret: Get \"https://localhost:6443/api/v1/namespaces/openshift-kube-controller-manager/secrets?limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate is valid for kubernetes, kubernetes.default, kubernetes.default.svc, kubernetes.default.svc.cluster.local, openshift, openshift.default, openshift.default.svc, openshift.default.svc.cluster.local, not localhost-recovery" 2025-12-04T11:55:45.587069990Z F1204 11:55:45.586929 1 base_controller.go:105] unable to sync caches for CertSyncController cluster-policy-controller_previous.log0000640000000000000000000000364015114274105036636 0ustar0000000000000000config/pod/openshift-kube-controller-manager/logs/kube-controller-manager-master-02025-12-04T12:01:28.412005903Z + timeout 3m /bin/bash -exuo pipefail -c 'while [ -n "$(ss -Htanop \( sport = 10357 \))" ]; do sleep 1; done' 2025-12-04T12:01:28.416748943Z ++ ss -Htanop '(' sport = 10357 ')' 2025-12-04T12:01:28.430581551Z + '[' -n '' ']' 2025-12-04T12:01:28.431564632Z + exec cluster-policy-controller start --config=/etc/kubernetes/static-pod-resources/configmaps/cluster-policy-controller-config/config.yaml --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager -v=2 2025-12-04T12:01:28.483117073Z I1204 12:01:28.482899 1 leaderelection.go:121] The leader election gives 4 retries and allows for 30s of clock skew. The kube-apiserver downtime tolerance is 78s. Worst non-graceful lease acquisition is 2m43s. Worst graceful lease acquisition is {26s}. 2025-12-04T12:01:28.483831916Z I1204 12:01:28.483760 1 observer_polling.go:159] Starting file observer 2025-12-04T12:01:28.484617021Z I1204 12:01:28.484564 1 builder.go:298] cluster-policy-controller version 4.18.0-202511181540.p2.g5fd8525.assembly.stream.el9-5fd8525-5fd852525909ce6eab52972ba9ce8fcf56528eb9 2025-12-04T12:01:28.485247621Z I1204 12:01:28.485189 1 dynamic_serving_content.go:116] "Loaded a new cert/key pair" name="serving-cert::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key" 2025-12-04T12:01:54.647387901Z I1204 12:01:54.647285 1 cmd.go:138] Received SIGTERM or SIGINT signal, shutting down controller. 2025-12-04T12:01:54.647438203Z F1204 12:01:54.647400 1 cmd.go:179] failed checking apiserver connectivity: Get "https://api-int.sno.openstack.lab:6443/apis/coordination.k8s.io/v1/namespaces/openshift-kube-controller-manager/leases/cluster-policy-controller-lock": tls: failed to verify certificate: x509: certificate signed by unknown authority config/pod/openshift-kube-controller-manager/kube-controller-manager-master-0.json0000640000000000000000000005021615114274105027047 0ustar0000000000000000{"metadata":{"name":"kube-controller-manager-master-0","namespace":"openshift-kube-controller-manager","uid":"8ba27afc-bec9-4bfc-b4a6-197a0564e7ba","resourceVersion":"14317","creationTimestamp":"2025-12-04T11:45:44Z","labels":{"app":"kube-controller-manager","kube-controller-manager":"true","revision":"2"},"annotations":{"kubectl.kubernetes.io/default-container":"kube-controller-manager","kubernetes.io/config.hash":"e6b437c60bb18680f4492b00b294e872","kubernetes.io/config.mirror":"e6b437c60bb18680f4492b00b294e872","kubernetes.io/config.seen":"2025-12-04T11:45:44.227874504Z","kubernetes.io/config.source":"file","target.workload.openshift.io/management":"{\"effect\": \"PreferredDuringScheduling\"}"},"ownerReferences":[{"apiVersion":"v1","kind":"Node","name":"master-0","uid":"9618c86a-8a87-4c00-8dd2-9747684d64b3","controller":true}]},"spec":{"volumes":[{"name":"resource-dir","hostPath":{"path":"/etc/kubernetes/static-pod-resources/kube-controller-manager-pod-2","type":""}},{"name":"cert-dir","hostPath":{"path":"/etc/kubernetes/static-pod-resources/kube-controller-manager-certs","type":""}}],"containers":[{"name":"kube-controller-manager","image":"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:6d5891cdd7dcf7c9081de8b364b4c96446b7f946f7880fbae291a4592a198264","command":["/bin/bash","-euxo","pipefail","-c"],"args":["timeout 3m /bin/bash -exuo pipefail -c 'while [ -n \"$(ss -Htanop \\( sport = 10257 \\))\" ]; do sleep 1; done'\n\nif [ -f /etc/kubernetes/static-pod-certs/configmaps/trusted-ca-bundle/ca-bundle.crt ]; then\n echo \"Copying system trust bundle\"\n cp -f /etc/kubernetes/static-pod-certs/configmaps/trusted-ca-bundle/ca-bundle.crt /etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem\nfi\n\nif [ -f /etc/kubernetes/static-pod-resources/configmaps/cloud-config/ca-bundle.pem ]; then\n echo \"Setting custom CA bundle for cloud provider\"\n export AWS_CA_BUNDLE=/etc/kubernetes/static-pod-resources/configmaps/cloud-config/ca-bundle.pem\nfi\n\nexec hyperkube kube-controller-manager --openshift-config=/etc/kubernetes/static-pod-resources/configmaps/config/config.yaml \\\n --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig \\\n --authentication-kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig \\\n --authorization-kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig \\\n --client-ca-file=/etc/kubernetes/static-pod-certs/configmaps/client-ca/ca-bundle.crt \\\n --requestheader-client-ca-file=/etc/kubernetes/static-pod-certs/configmaps/aggregator-client-ca/ca-bundle.crt -v=2 --allocate-node-cidrs=false --cert-dir=/var/run/kubernetes --cloud-provider=external --cluster-cidr=10.128.0.0/16 --cluster-name=sno-d7v9r --cluster-signing-cert-file=/etc/kubernetes/static-pod-certs/secrets/csr-signer/tls.crt --cluster-signing-duration=720h --cluster-signing-key-file=/etc/kubernetes/static-pod-certs/secrets/csr-signer/tls.key --controllers=* --controllers=-bootstrapsigner --controllers=-tokencleaner --controllers=-ttl --enable-dynamic-provisioning=true --feature-gates=AWSClusterHostedDNS=false --feature-gates=AWSEFSDriverVolumeMetrics=true --feature-gates=AdditionalRoutingCapabilities=false --feature-gates=AdminNetworkPolicy=true --feature-gates=AlibabaPlatform=true --feature-gates=AutomatedEtcdBackup=false --feature-gates=AzureWorkloadIdentity=true --feature-gates=BareMetalLoadBalancer=true --feature-gates=BootcNodeManagement=false --feature-gates=BuildCSIVolumes=true --feature-gates=CSIDriverSharedResource=false --feature-gates=ChunkSizeMiB=true --feature-gates=CloudDualStackNodeIPs=true --feature-gates=ClusterAPIInstall=false --feature-gates=ClusterAPIInstallIBMCloud=false --feature-gates=ClusterMonitoringConfig=false --feature-gates=ConsolePluginContentSecurityPolicy=false --feature-gates=DNSNameResolver=false --feature-gates=DisableKubeletCloudCredentialProviders=true --feature-gates=DynamicResourceAllocation=false --feature-gates=EtcdBackendQuota=false --feature-gates=EventedPLEG=false --feature-gates=Example=false --feature-gates=ExternalOIDC=false --feature-gates=ExternalOIDCWithUIDAndExtraClaimMappings=false --feature-gates=GCPClusterHostedDNS=false --feature-gates=GCPLabelsTags=true --feature-gates=GatewayAPI=false --feature-gates=HardwareSpeed=true --feature-gates=ImageStreamImportMode=false --feature-gates=IngressControllerDynamicConfigurationManager=false --feature-gates=IngressControllerLBSubnetsAWS=true --feature-gates=InsightsConfig=false --feature-gates=InsightsConfigAPI=false --feature-gates=InsightsOnDemandDataGather=false --feature-gates=InsightsRuntimeExtractor=false --feature-gates=KMSv1=true --feature-gates=MachineAPIMigration=false --feature-gates=MachineAPIOperatorDisableMachineHealthCheckController=false --feature-gates=MachineAPIProviderOpenStack=false --feature-gates=MachineConfigNodes=false --feature-gates=ManagedBootImages=true --feature-gates=ManagedBootImagesAWS=true --feature-gates=MaxUnavailableStatefulSet=false --feature-gates=MetricsCollectionProfiles=false --feature-gates=MinimumKubeletVersion=false --feature-gates=MixedCPUsAllocation=false --feature-gates=MultiArchInstallAWS=true --feature-gates=MultiArchInstallAzure=false --feature-gates=MultiArchInstallGCP=true --feature-gates=NetworkDiagnosticsConfig=true --feature-gates=NetworkLiveMigration=true --feature-gates=NetworkSegmentation=true --feature-gates=NewOLM=true --feature-gates=NodeDisruptionPolicy=true --feature-gates=NodeSwap=false --feature-gates=NutanixMultiSubnets=false --feature-gates=OVNObservability=false --feature-gates=OnClusterBuild=true --feature-gates=OpenShiftPodSecurityAdmission=false --feature-gates=PersistentIPsForVirtualization=true --feature-gates=PinnedImages=false --feature-gates=PlatformOperators=false --feature-gates=PrivateHostedZoneAWS=true --feature-gates=ProcMountType=false --feature-gates=RouteAdvertisements=false --feature-gates=RouteExternalCertificate=false --feature-gates=ServiceAccountTokenNodeBinding=false --feature-gates=SetEIPForNLBIngressController=true --feature-gates=SignatureStores=false --feature-gates=SigstoreImageVerification=false --feature-gates=StreamingCollectionEncodingToJSON=false --feature-gates=StreamingCollectionEncodingToProtobuf=false --feature-gates=TranslateStreamCloseWebsocketRequests=false --feature-gates=UpgradeStatus=false --feature-gates=UserNamespacesPodSecurityStandards=false --feature-gates=UserNamespacesSupport=false --feature-gates=VSphereControlPlaneMachineSet=true --feature-gates=VSphereDriverConfiguration=true --feature-gates=VSphereMultiNetworks=false --feature-gates=VSphereMultiVCenters=true --feature-gates=VSphereStaticIPs=true --feature-gates=ValidatingAdmissionPolicy=true --feature-gates=VolumeAttributesClass=false --feature-gates=VolumeGroupSnapshot=false --flex-volume-plugin-dir=/etc/kubernetes/kubelet-plugins/volume/exec --kube-api-burst=300 --kube-api-qps=150 --leader-elect-renew-deadline=12s --leader-elect-resource-lock=leases --leader-elect-retry-period=3s --leader-elect=true --pv-recycler-pod-template-filepath-hostpath=/etc/kubernetes/static-pod-resources/configmaps/recycler-config/recycler-pod.yaml --pv-recycler-pod-template-filepath-nfs=/etc/kubernetes/static-pod-resources/configmaps/recycler-config/recycler-pod.yaml --root-ca-file=/etc/kubernetes/static-pod-resources/configmaps/serviceaccount-ca/ca-bundle.crt --secure-port=10257 --service-account-private-key-file=/etc/kubernetes/static-pod-resources/secrets/service-account-private-key/service-account.key --service-cluster-ip-range=172.30.0.0/16 --use-service-account-credentials=true --tls-cipher-suites=TLS_AES_128_GCM_SHA256,TLS_AES_256_GCM_SHA384,TLS_CHACHA20_POLY1305_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305_SHA256,TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305_SHA256 --tls-min-version=VersionTLS12"],"ports":[{"hostPort":10257,"containerPort":10257,"protocol":"TCP"}],"resources":{"requests":{"cpu":"60m","memory":"200Mi"}},"volumeMounts":[{"name":"resource-dir","mountPath":"/etc/kubernetes/static-pod-resources"},{"name":"cert-dir","mountPath":"/etc/kubernetes/static-pod-certs"}],"livenessProbe":{"httpGet":{"path":"healthz","port":10257,"scheme":"HTTPS"},"initialDelaySeconds":45,"timeoutSeconds":10,"periodSeconds":10,"successThreshold":1,"failureThreshold":3},"readinessProbe":{"httpGet":{"path":"healthz","port":10257,"scheme":"HTTPS"},"initialDelaySeconds":10,"timeoutSeconds":10,"periodSeconds":10,"successThreshold":1,"failureThreshold":3},"startupProbe":{"httpGet":{"path":"healthz","port":10257,"scheme":"HTTPS"},"timeoutSeconds":3,"periodSeconds":10,"successThreshold":1,"failureThreshold":3},"terminationMessagePath":"/dev/termination-log","terminationMessagePolicy":"FallbackToLogsOnError","imagePullPolicy":"IfNotPresent"},{"name":"cluster-policy-controller","image":"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d64c13fe7663a0b4ae61d103b1b7598adcf317a01826f296bcb66b1a2de83c96","command":["/bin/bash","-euxo","pipefail","-c"],"args":["timeout 3m /bin/bash -exuo pipefail -c 'while [ -n \"$(ss -Htanop \\( sport = 10357 \\))\" ]; do sleep 1; done'\n\nexec cluster-policy-controller start --config=/etc/kubernetes/static-pod-resources/configmaps/cluster-policy-controller-config/config.yaml \\\n --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig \\\n --namespace=${POD_NAMESPACE} -v=2"],"ports":[{"hostPort":10357,"containerPort":10357,"protocol":"TCP"}],"env":[{"name":"POD_NAME","valueFrom":{"fieldRef":{"apiVersion":"v1","fieldPath":"metadata.name"}}},{"name":"POD_NAMESPACE","valueFrom":{"fieldRef":{"apiVersion":"v1","fieldPath":"metadata.namespace"}}}],"resources":{"requests":{"cpu":"10m","memory":"200Mi"}},"volumeMounts":[{"name":"resource-dir","mountPath":"/etc/kubernetes/static-pod-resources"},{"name":"cert-dir","mountPath":"/etc/kubernetes/static-pod-certs"}],"livenessProbe":{"httpGet":{"path":"healthz","port":10357,"host":"localhost","scheme":"HTTPS"},"initialDelaySeconds":45,"timeoutSeconds":10,"periodSeconds":10,"successThreshold":1,"failureThreshold":3},"readinessProbe":{"httpGet":{"path":"healthz","port":10357,"host":"localhost","scheme":"HTTPS"},"initialDelaySeconds":10,"timeoutSeconds":10,"periodSeconds":10,"successThreshold":1,"failureThreshold":3},"startupProbe":{"httpGet":{"path":"healthz","port":10357,"host":"localhost","scheme":"HTTPS"},"timeoutSeconds":3,"periodSeconds":10,"successThreshold":1,"failureThreshold":3},"terminationMessagePath":"/dev/termination-log","terminationMessagePolicy":"FallbackToLogsOnError","imagePullPolicy":"IfNotPresent"},{"name":"kube-controller-manager-cert-syncer","image":"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a824e468cf8dd61d347e35b2ee5bc2f815666957647098e21a1bb56ff613e5b9","command":["cluster-kube-controller-manager-operator","cert-syncer"],"args":["--kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/kube-controller-cert-syncer-kubeconfig/kubeconfig","--namespace=$(POD_NAMESPACE)","--destination-dir=/etc/kubernetes/static-pod-certs"],"env":[{"name":"POD_NAME","valueFrom":{"fieldRef":{"apiVersion":"v1","fieldPath":"metadata.name"}}},{"name":"POD_NAMESPACE","valueFrom":{"fieldRef":{"apiVersion":"v1","fieldPath":"metadata.namespace"}}}],"resources":{"requests":{"cpu":"5m","memory":"50Mi"}},"volumeMounts":[{"name":"resource-dir","mountPath":"/etc/kubernetes/static-pod-resources"},{"name":"cert-dir","mountPath":"/etc/kubernetes/static-pod-certs"}],"terminationMessagePath":"/dev/termination-log","terminationMessagePolicy":"FallbackToLogsOnError","imagePullPolicy":"IfNotPresent"},{"name":"kube-controller-manager-recovery-controller","image":"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a824e468cf8dd61d347e35b2ee5bc2f815666957647098e21a1bb56ff613e5b9","command":["/bin/bash","-euxo","pipefail","-c"],"args":["timeout 3m /bin/bash -exuo pipefail -c 'while [ -n \"$(ss -Htanop \\( sport = 9443 \\))\" ]; do sleep 1; done'\n\nexec cluster-kube-controller-manager-operator cert-recovery-controller --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/kube-controller-cert-syncer-kubeconfig/kubeconfig --namespace=${POD_NAMESPACE} --listen=0.0.0.0:9443 -v=2"],"env":[{"name":"POD_NAMESPACE","valueFrom":{"fieldRef":{"apiVersion":"v1","fieldPath":"metadata.namespace"}}}],"resources":{"requests":{"cpu":"5m","memory":"50Mi"}},"volumeMounts":[{"name":"resource-dir","mountPath":"/etc/kubernetes/static-pod-resources"},{"name":"cert-dir","mountPath":"/etc/kubernetes/static-pod-certs"}],"terminationMessagePath":"/dev/termination-log","terminationMessagePolicy":"FallbackToLogsOnError","imagePullPolicy":"IfNotPresent"}],"restartPolicy":"Always","terminationGracePeriodSeconds":30,"dnsPolicy":"ClusterFirst","nodeName":"master-0","hostNetwork":true,"securityContext":{},"schedulerName":"default-scheduler","tolerations":[{"operator":"Exists"}],"priorityClassName":"system-node-critical","priority":2000001000,"enableServiceLinks":true,"preemptionPolicy":"PreemptLowerPriority"},"status":{"phase":"Running","conditions":[{"type":"PodReadyToStartContainers","status":"True","lastProbeTime":null,"lastTransitionTime":"2025-12-04T11:59:18Z"},{"type":"Initialized","status":"True","lastProbeTime":null,"lastTransitionTime":"2025-12-04T11:59:18Z"},{"type":"Ready","status":"False","lastProbeTime":null,"lastTransitionTime":"2025-12-04T11:59:18Z","reason":"ContainersNotReady","message":"containers with unready status: [cluster-policy-controller]"},{"type":"ContainersReady","status":"False","lastProbeTime":null,"lastTransitionTime":"2025-12-04T11:59:18Z","reason":"ContainersNotReady","message":"containers with unready status: [cluster-policy-controller]"},{"type":"PodScheduled","status":"True","lastProbeTime":null,"lastTransitionTime":"2025-12-04T11:59:18Z"}],"hostIP":"192.168.32.10","hostIPs":[{"ip":"192.168.32.10"}],"podIP":"192.168.32.10","podIPs":[{"ip":"192.168.32.10"}],"startTime":"2025-12-04T11:59:18Z","containerStatuses":[{"name":"cluster-policy-controller","state":{"waiting":{"reason":"CrashLoopBackOff","message":"back-off 1m20s restarting failed container=cluster-policy-controller pod=kube-controller-manager-master-0_openshift-kube-controller-manager(e6b437c60bb18680f4492b00b294e872)"}},"lastState":{"terminated":{"exitCode":255,"reason":"Error","message":"+ timeout 3m /bin/bash -exuo pipefail -c 'while [ -n \"$(ss -Htanop \\( sport = 10357 \\))\" ]; do sleep 1; done'\n++ ss -Htanop '(' sport = 10357 ')'\n+ '[' -n '' ']'\n+ exec cluster-policy-controller start --config=/etc/kubernetes/static-pod-resources/configmaps/cluster-policy-controller-config/config.yaml --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager -v=2\nI1204 12:01:28.482899 1 leaderelection.go:121] The leader election gives 4 retries and allows for 30s of clock skew. The kube-apiserver downtime tolerance is 78s. Worst non-graceful lease acquisition is 2m43s. Worst graceful lease acquisition is {26s}.\nI1204 12:01:28.483760 1 observer_polling.go:159] Starting file observer\nI1204 12:01:28.484564 1 builder.go:298] cluster-policy-controller version 4.18.0-202511181540.p2.g5fd8525.assembly.stream.el9-5fd8525-5fd852525909ce6eab52972ba9ce8fcf56528eb9\nI1204 12:01:28.485189 1 dynamic_serving_content.go:116] \"Loaded a new cert/key pair\" name=\"serving-cert::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key\"\nI1204 12:01:54.647285 1 cmd.go:138] Received SIGTERM or SIGINT signal, shutting down controller.\nF1204 12:01:54.647400 1 cmd.go:179] failed checking apiserver connectivity: Get \"https://api-int.sno.openstack.lab:6443/apis/coordination.k8s.io/v1/namespaces/openshift-kube-controller-manager/leases/cluster-policy-controller-lock\": tls: failed to verify certificate: x509: certificate signed by unknown authority\n","startedAt":"2025-12-04T12:01:28Z","finishedAt":"2025-12-04T12:01:54Z","containerID":"cri-o://f4af8a14216534dca63f05d24b76a32e74c089da1cc06d3e232ed5d109ed4493"}},"ready":false,"restartCount":11,"image":"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d64c13fe7663a0b4ae61d103b1b7598adcf317a01826f296bcb66b1a2de83c96","imageID":"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d64c13fe7663a0b4ae61d103b1b7598adcf317a01826f296bcb66b1a2de83c96","containerID":"cri-o://f4af8a14216534dca63f05d24b76a32e74c089da1cc06d3e232ed5d109ed4493","started":false},{"name":"kube-controller-manager","state":{"running":{"startedAt":"2025-12-04T11:45:44Z"}},"lastState":{},"ready":true,"restartCount":0,"image":"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:6d5891cdd7dcf7c9081de8b364b4c96446b7f946f7880fbae291a4592a198264","imageID":"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:6d5891cdd7dcf7c9081de8b364b4c96446b7f946f7880fbae291a4592a198264","containerID":"cri-o://f78a7d287972772a3c979bede114291f869074b909122bce68dcbdb34f2d707c","started":true},{"name":"kube-controller-manager-cert-syncer","state":{"running":{"startedAt":"2025-12-04T11:55:46Z"}},"lastState":{"terminated":{"exitCode":1,"reason":"Error","message":" failed to verify certificate: x509: certificate is valid for kubernetes, kubernetes.default, kubernetes.default.svc, kubernetes.default.svc.cluster.local, openshift, openshift.default, openshift.default.svc, openshift.default.svc.cluster.local, not localhost-recovery\nE1204 11:55:00.404388 1 reflector.go:158] \"Unhandled Error\" err=\"k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243: Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: Get \\\"https://localhost:6443/api/v1/namespaces/openshift-kube-controller-manager/configmaps?limit=500\u0026resourceVersion=0\\\": tls: failed to verify certificate: x509: certificate is valid for kubernetes, kubernetes.default, kubernetes.default.svc, kubernetes.default.svc.cluster.local, openshift, openshift.default, openshift.default.svc, openshift.default.svc.cluster.local, not localhost-recovery\"\nW1204 11:55:40.529780 1 reflector.go:561] k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243: failed to list *v1.Secret: Get \"https://localhost:6443/api/v1/namespaces/openshift-kube-controller-manager/secrets?limit=500\u0026resourceVersion=0\": tls: failed to verify certificate: x509: certificate is valid for kubernetes, kubernetes.default, kubernetes.default.svc, kubernetes.default.svc.cluster.local, openshift, openshift.default, openshift.default.svc, openshift.default.svc.cluster.local, not localhost-recovery\nE1204 11:55:40.529981 1 reflector.go:158] \"Unhandled Error\" err=\"k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243: Failed to watch *v1.Secret: failed to list *v1.Secret: Get \\\"https://localhost:6443/api/v1/namespaces/openshift-kube-controller-manager/secrets?limit=500\u0026resourceVersion=0\\\": tls: failed to verify certificate: x509: certificate is valid for kubernetes, kubernetes.default, kubernetes.default.svc, kubernetes.default.svc.cluster.local, openshift, openshift.default, openshift.default.svc, openshift.default.svc.cluster.local, not localhost-recovery\"\nF1204 11:55:45.586929 1 base_controller.go:105] unable to sync caches for CertSyncController\n","startedAt":"2025-12-04T11:45:45Z","finishedAt":"2025-12-04T11:55:45Z","containerID":"cri-o://11e29c5651a5425702af55ec3b8ee473df3972ab82e099ae3490c6b27969c94a"}},"ready":true,"restartCount":1,"image":"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a824e468cf8dd61d347e35b2ee5bc2f815666957647098e21a1bb56ff613e5b9","imageID":"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a824e468cf8dd61d347e35b2ee5bc2f815666957647098e21a1bb56ff613e5b9","containerID":"cri-o://c7daeda85ec183c8d113e00834a6cc8eb19542aa3278f40e44cf98acd7483141","started":true},{"name":"kube-controller-manager-recovery-controller","state":{"running":{"startedAt":"2025-12-04T11:45:45Z"}},"lastState":{},"ready":true,"restartCount":0,"image":"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a824e468cf8dd61d347e35b2ee5bc2f815666957647098e21a1bb56ff613e5b9","imageID":"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a824e468cf8dd61d347e35b2ee5bc2f815666957647098e21a1bb56ff613e5b9","containerID":"cri-o://a4300d8490d9cbb0f00793e382d3ed2318a1bd59a39a4b37b7e8bc4be83371b7","started":true}],"qosClass":"Burstable"}}config/pod/openshift-kube-controller-manager/logs/installer-3-master-0/installer_current.log0000640000000000000000000001443515114274105031003 0ustar00000000000000002025-12-04T11:50:26.400953598Z I1204 11:50:26.400639 1 cmd.go:95] &{ true {false} installer true map[cert-configmaps:0xc0008b6000 cert-dir:0xc0008b61e0 cert-secrets:0xc000625ea0 configmaps:0xc000625a40 namespace:0xc000625860 optional-cert-configmaps:0xc0008b6140 optional-configmaps:0xc000625b80 optional-secrets:0xc000625ae0 pod:0xc000625900 pod-manifest-dir:0xc000625cc0 resource-dir:0xc000625c20 revision:0xc0006257c0 secrets:0xc0006259a0 v:0xc0008b6be0] [0xc0008b6be0 0xc0006257c0 0xc000625860 0xc000625900 0xc000625c20 0xc000625cc0 0xc000625a40 0xc000625b80 0xc0006259a0 0xc000625ae0 0xc0008b61e0 0xc0008b6000 0xc0008b6140 0xc000625ea0] [] map[cert-configmaps:0xc0008b6000 cert-dir:0xc0008b61e0 cert-secrets:0xc000625ea0 configmaps:0xc000625a40 help:0xc0008b6fa0 kubeconfig:0xc000625720 log-flush-frequency:0xc0008b6b40 namespace:0xc000625860 optional-cert-configmaps:0xc0008b6140 optional-cert-secrets:0xc0008b60a0 optional-configmaps:0xc000625b80 optional-secrets:0xc000625ae0 pod:0xc000625900 pod-manifest-dir:0xc000625cc0 pod-manifests-lock-file:0xc000625e00 resource-dir:0xc000625c20 revision:0xc0006257c0 secrets:0xc0006259a0 timeout-duration:0xc000625d60 v:0xc0008b6be0 vmodule:0xc0008b6c80] [0xc000625720 0xc0006257c0 0xc000625860 0xc000625900 0xc0006259a0 0xc000625a40 0xc000625ae0 0xc000625b80 0xc000625c20 0xc000625cc0 0xc000625d60 0xc000625e00 0xc000625ea0 0xc0008b6000 0xc0008b60a0 0xc0008b6140 0xc0008b61e0 0xc0008b6b40 0xc0008b6be0 0xc0008b6c80 0xc0008b6fa0] [0xc0008b6000 0xc0008b61e0 0xc000625ea0 0xc000625a40 0xc0008b6fa0 0xc000625720 0xc0008b6b40 0xc000625860 0xc0008b6140 0xc0008b60a0 0xc000625b80 0xc000625ae0 0xc000625900 0xc000625cc0 0xc000625e00 0xc000625c20 0xc0006257c0 0xc0006259a0 0xc000625d60 0xc0008b6be0 0xc0008b6c80] map[104:0xc0008b6fa0 118:0xc0008b6be0] [] -1 0 0xc0008ac030 true 0x77b500 []} 2025-12-04T11:50:26.401064032Z I1204 11:50:26.400961 1 cmd.go:96] (*installerpod.InstallOptions)(0xc000049520)({ 2025-12-04T11:50:26.401064032Z KubeConfig: (string) "", 2025-12-04T11:50:26.401064032Z KubeClient: (kubernetes.Interface) , 2025-12-04T11:50:26.401064032Z Revision: (string) (len=1) "3", 2025-12-04T11:50:26.401064032Z NodeName: (string) "", 2025-12-04T11:50:26.401064032Z Namespace: (string) (len=33) "openshift-kube-controller-manager", 2025-12-04T11:50:26.401064032Z Clock: (clock.RealClock) { 2025-12-04T11:50:26.401064032Z }, 2025-12-04T11:50:26.401064032Z PodConfigMapNamePrefix: (string) (len=27) "kube-controller-manager-pod", 2025-12-04T11:50:26.401064032Z SecretNamePrefixes: ([]string) (len=2 cap=2) { 2025-12-04T11:50:26.401064032Z (string) (len=27) "service-account-private-key", 2025-12-04T11:50:26.401064032Z (string) (len=31) "localhost-recovery-client-token" 2025-12-04T11:50:26.401064032Z }, 2025-12-04T11:50:26.401064032Z OptionalSecretNamePrefixes: ([]string) (len=1 cap=1) { 2025-12-04T11:50:26.401064032Z (string) (len=12) "serving-cert" 2025-12-04T11:50:26.401064032Z }, 2025-12-04T11:50:26.401064032Z ConfigMapNamePrefixes: ([]string) (len=8 cap=8) { 2025-12-04T11:50:26.401064032Z (string) (len=27) "kube-controller-manager-pod", 2025-12-04T11:50:26.401064032Z (string) (len=6) "config", 2025-12-04T11:50:26.401064032Z (string) (len=32) "cluster-policy-controller-config", 2025-12-04T11:50:26.401064032Z (string) (len=29) "controller-manager-kubeconfig", 2025-12-04T11:50:26.401064032Z (string) (len=38) "kube-controller-cert-syncer-kubeconfig", 2025-12-04T11:50:26.401064032Z (string) (len=17) "serviceaccount-ca", 2025-12-04T11:50:26.401064032Z (string) (len=10) "service-ca", 2025-12-04T11:50:26.401064032Z (string) (len=15) "recycler-config" 2025-12-04T11:50:26.401064032Z }, 2025-12-04T11:50:26.401064032Z OptionalConfigMapNamePrefixes: ([]string) (len=1 cap=1) { 2025-12-04T11:50:26.401064032Z (string) (len=12) "cloud-config" 2025-12-04T11:50:26.401064032Z }, 2025-12-04T11:50:26.401064032Z CertSecretNames: ([]string) (len=2 cap=2) { 2025-12-04T11:50:26.401064032Z (string) (len=39) "kube-controller-manager-client-cert-key", 2025-12-04T11:50:26.401064032Z (string) (len=10) "csr-signer" 2025-12-04T11:50:26.401064032Z }, 2025-12-04T11:50:26.401064032Z OptionalCertSecretNamePrefixes: ([]string) , 2025-12-04T11:50:26.401064032Z CertConfigMapNamePrefixes: ([]string) (len=2 cap=2) { 2025-12-04T11:50:26.401064032Z (string) (len=20) "aggregator-client-ca", 2025-12-04T11:50:26.401064032Z (string) (len=9) "client-ca" 2025-12-04T11:50:26.401064032Z }, 2025-12-04T11:50:26.401064032Z OptionalCertConfigMapNamePrefixes: ([]string) (len=1 cap=1) { 2025-12-04T11:50:26.401064032Z (string) (len=17) "trusted-ca-bundle" 2025-12-04T11:50:26.401064032Z }, 2025-12-04T11:50:26.401064032Z CertDir: (string) (len=66) "/etc/kubernetes/static-pod-resources/kube-controller-manager-certs", 2025-12-04T11:50:26.401064032Z ResourceDir: (string) (len=36) "/etc/kubernetes/static-pod-resources", 2025-12-04T11:50:26.401064032Z PodManifestDir: (string) (len=25) "/etc/kubernetes/manifests", 2025-12-04T11:50:26.401064032Z Timeout: (time.Duration) 2m0s, 2025-12-04T11:50:26.401064032Z StaticPodManifestsLockFile: (string) "", 2025-12-04T11:50:26.401064032Z PodMutationFns: ([]installerpod.PodMutationFunc) , 2025-12-04T11:50:26.401064032Z KubeletVersion: (string) "" 2025-12-04T11:50:26.401064032Z }) 2025-12-04T11:50:26.401842454Z I1204 11:50:26.401782 1 cmd.go:413] Getting controller reference for node master-0 2025-12-04T11:50:26.499341807Z I1204 11:50:26.499253 1 cmd.go:426] Waiting for installer revisions to settle for node master-0 2025-12-04T11:50:26.499341807Z I1204 11:50:26.499327 1 envvar.go:172] "Feature gate default state" feature="WatchListClient" enabled=false 2025-12-04T11:50:26.499395689Z I1204 11:50:26.499338 1 envvar.go:172] "Feature gate default state" feature="InformerResourceVersion" enabled=false 2025-12-04T11:50:26.502020458Z I1204 11:50:26.501970 1 cmd.go:518] Waiting additional period after revisions have settled for node master-0 2025-12-04T11:50:56.502515904Z I1204 11:50:56.502405 1 cmd.go:524] Getting installer pods for node master-0 2025-12-04T11:51:10.506407317Z F1204 11:51:10.506236 1 cmd.go:109] Get "https://172.30.0.1:443/api/v1/namespaces/openshift-kube-controller-manager/pods?labelSelector=app%3Dinstaller": net/http: request canceled (Client.Timeout exceeded while awaiting headers) config/pod/openshift-kube-controller-manager/installer-3-master-0.json0000640000000000000000000001717115114274105024470 0ustar0000000000000000{"metadata":{"name":"installer-3-master-0","namespace":"openshift-kube-controller-manager","uid":"27c4e782-11b4-43a3-9ffd-34bf4e40bbc4","resourceVersion":"11904","creationTimestamp":"2025-12-04T11:50:24Z","labels":{"app":"installer"},"annotations":{"k8s.ovn.org/pod-networks":"{\"default\":{\"ip_addresses\":[\"10.128.0.73/23\"],\"mac_address\":\"0a:58:0a:80:00:49\",\"gateway_ips\":[\"10.128.0.1\"],\"routes\":[{\"dest\":\"10.128.0.0/16\",\"nextHop\":\"10.128.0.1\"},{\"dest\":\"172.30.0.0/16\",\"nextHop\":\"10.128.0.1\"},{\"dest\":\"169.254.0.5/32\",\"nextHop\":\"10.128.0.1\"},{\"dest\":\"100.64.0.0/16\",\"nextHop\":\"10.128.0.1\"}],\"ip_address\":\"10.128.0.73/23\",\"gateway_ip\":\"10.128.0.1\",\"role\":\"primary\"}}","k8s.v1.cni.cncf.io/network-status":"[{\n \"name\": \"ovn-kubernetes\",\n \"interface\": \"eth0\",\n \"ips\": [\n \"10.128.0.73\"\n ],\n \"mac\": \"0a:58:0a:80:00:49\",\n \"default\": true,\n \"dns\": {}\n}]"},"ownerReferences":[{"apiVersion":"v1","kind":"ConfigMap","name":"revision-status-3","uid":"c3000b30-b4db-42cf-8f60-b9d8f9ea9787"}]},"spec":{"volumes":[{"name":"kubelet-dir","hostPath":{"path":"/etc/kubernetes/","type":""}},{"name":"var-lock","hostPath":{"path":"/var/lock","type":""}},{"name":"kube-api-access","projected":{"sources":[{"serviceAccountToken":{"expirationSeconds":3600,"path":"token"}},{"configMap":{"name":"kube-root-ca.crt","items":[{"key":"ca.crt","path":"ca.crt"}]}},{"downwardAPI":{"items":[{"path":"namespace","fieldRef":{"apiVersion":"v1","fieldPath":"metadata.namespace"}}]}}],"defaultMode":420}}],"containers":[{"name":"installer","image":"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a824e468cf8dd61d347e35b2ee5bc2f815666957647098e21a1bb56ff613e5b9","command":["cluster-kube-controller-manager-operator","installer"],"args":["-v=2","--revision=3","--namespace=openshift-kube-controller-manager","--pod=kube-controller-manager-pod","--resource-dir=/etc/kubernetes/static-pod-resources","--pod-manifest-dir=/etc/kubernetes/manifests","--configmaps=kube-controller-manager-pod","--configmaps=config","--configmaps=cluster-policy-controller-config","--configmaps=controller-manager-kubeconfig","--optional-configmaps=cloud-config","--configmaps=kube-controller-cert-syncer-kubeconfig","--configmaps=serviceaccount-ca","--configmaps=service-ca","--configmaps=recycler-config","--secrets=service-account-private-key","--optional-secrets=serving-cert","--secrets=localhost-recovery-client-token","--cert-dir=/etc/kubernetes/static-pod-resources/kube-controller-manager-certs","--cert-configmaps=aggregator-client-ca","--cert-configmaps=client-ca","--optional-cert-configmaps=trusted-ca-bundle","--cert-secrets=kube-controller-manager-client-cert-key","--cert-secrets=csr-signer"],"env":[{"name":"POD_NAME","valueFrom":{"fieldRef":{"apiVersion":"v1","fieldPath":"metadata.name"}}},{"name":"NODE_NAME","valueFrom":{"fieldRef":{"apiVersion":"v1","fieldPath":"spec.nodeName"}}}],"resources":{"limits":{"cpu":"150m","memory":"200M"},"requests":{"cpu":"150m","memory":"200M"}},"volumeMounts":[{"name":"kubelet-dir","mountPath":"/etc/kubernetes/"},{"name":"kube-api-access","readOnly":true,"mountPath":"/var/run/secrets/kubernetes.io/serviceaccount"},{"name":"var-lock","mountPath":"/var/lock"}],"terminationMessagePath":"/dev/termination-log","terminationMessagePolicy":"FallbackToLogsOnError","imagePullPolicy":"IfNotPresent","securityContext":{"privileged":true,"runAsUser":0}}],"restartPolicy":"Never","terminationGracePeriodSeconds":30,"dnsPolicy":"ClusterFirst","serviceAccountName":"installer-sa","serviceAccount":"installer-sa","automountServiceAccountToken":false,"nodeName":"master-0","securityContext":{"runAsUser":0},"imagePullSecrets":[{"name":"installer-sa-dockercfg-vn92f"}],"schedulerName":"default-scheduler","tolerations":[{"operator":"Exists"}],"priorityClassName":"system-node-critical","priority":2000001000,"enableServiceLinks":true,"preemptionPolicy":"PreemptLowerPriority"},"status":{"phase":"Failed","conditions":[{"type":"PodReadyToStartContainers","status":"False","lastProbeTime":null,"lastTransitionTime":"2025-12-04T11:51:12Z"},{"type":"Initialized","status":"True","lastProbeTime":null,"lastTransitionTime":"2025-12-04T11:50:24Z"},{"type":"Ready","status":"False","lastProbeTime":null,"lastTransitionTime":"2025-12-04T11:51:10Z","reason":"PodFailed"},{"type":"ContainersReady","status":"False","lastProbeTime":null,"lastTransitionTime":"2025-12-04T11:51:10Z","reason":"PodFailed"},{"type":"PodScheduled","status":"True","lastProbeTime":null,"lastTransitionTime":"2025-12-04T11:50:24Z"}],"hostIP":"192.168.32.10","hostIPs":[{"ip":"192.168.32.10"}],"startTime":"2025-12-04T11:50:24Z","containerStatuses":[{"name":"installer","state":{"terminated":{"exitCode":1,"reason":"Error","message":"icy-controller-config\",\n (string) (len=29) \"controller-manager-kubeconfig\",\n (string) (len=38) \"kube-controller-cert-syncer-kubeconfig\",\n (string) (len=17) \"serviceaccount-ca\",\n (string) (len=10) \"service-ca\",\n (string) (len=15) \"recycler-config\"\n },\n OptionalConfigMapNamePrefixes: ([]string) (len=1 cap=1) {\n (string) (len=12) \"cloud-config\"\n },\n CertSecretNames: ([]string) (len=2 cap=2) {\n (string) (len=39) \"kube-controller-manager-client-cert-key\",\n (string) (len=10) \"csr-signer\"\n },\n OptionalCertSecretNamePrefixes: ([]string) \u003cnil\u003e,\n CertConfigMapNamePrefixes: ([]string) (len=2 cap=2) {\n (string) (len=20) \"aggregator-client-ca\",\n (string) (len=9) \"client-ca\"\n },\n OptionalCertConfigMapNamePrefixes: ([]string) (len=1 cap=1) {\n (string) (len=17) \"trusted-ca-bundle\"\n },\n CertDir: (string) (len=66) \"/etc/kubernetes/static-pod-resources/kube-controller-manager-certs\",\n ResourceDir: (string) (len=36) \"/etc/kubernetes/static-pod-resources\",\n PodManifestDir: (string) (len=25) \"/etc/kubernetes/manifests\",\n Timeout: (time.Duration) 2m0s,\n StaticPodManifestsLockFile: (string) \"\",\n PodMutationFns: ([]installerpod.PodMutationFunc) \u003cnil\u003e,\n KubeletVersion: (string) \"\"\n})\nI1204 11:50:26.401782 1 cmd.go:413] Getting controller reference for node master-0\nI1204 11:50:26.499253 1 cmd.go:426] Waiting for installer revisions to settle for node master-0\nI1204 11:50:26.499327 1 envvar.go:172] \"Feature gate default state\" feature=\"WatchListClient\" enabled=false\nI1204 11:50:26.499338 1 envvar.go:172] \"Feature gate default state\" feature=\"InformerResourceVersion\" enabled=false\nI1204 11:50:26.501970 1 cmd.go:518] Waiting additional period after revisions have settled for node master-0\nI1204 11:50:56.502405 1 cmd.go:524] Getting installer pods for node master-0\nF1204 11:51:10.506236 1 cmd.go:109] Get \"https://172.30.0.1:443/api/v1/namespaces/openshift-kube-controller-manager/pods?labelSelector=app%3Dinstaller\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)\n","startedAt":"2025-12-04T11:50:25Z","finishedAt":"2025-12-04T11:51:10Z","containerID":"cri-o://b78e17b7306f7adda3e6d7075255514653aaae6ce42d92ca0bc99707e1589503"}},"lastState":{},"ready":false,"restartCount":0,"image":"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a824e468cf8dd61d347e35b2ee5bc2f815666957647098e21a1bb56ff613e5b9","imageID":"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a824e468cf8dd61d347e35b2ee5bc2f815666957647098e21a1bb56ff613e5b9","containerID":"cri-o://b78e17b7306f7adda3e6d7075255514653aaae6ce42d92ca0bc99707e1589503","started":false,"volumeMounts":[{"name":"kubelet-dir","mountPath":"/etc/kubernetes/"},{"name":"kube-api-access","mountPath":"/var/run/secrets/kubernetes.io/serviceaccount","readOnly":true,"recursiveReadOnly":"Disabled"},{"name":"var-lock","mountPath":"/var/lock"}]}],"qosClass":"Guaranteed"}}config/pod/openshift-kube-controller-manager/logs/installer-2-master-0/installer_current.log0000640000000000000000000011332715114274105031002 0ustar00000000000000002025-12-04T11:45:12.196832575Z (string) (len=10) "csr-signer" 2025-12-04T11:45:12.196832575Z }, 2025-12-04T11:45:12.196832575Z OptionalCertSecretNamePrefixes: ([]string) , 2025-12-04T11:45:12.196832575Z CertConfigMapNamePrefixes: ([]string) (len=2 cap=2) { 2025-12-04T11:45:12.196832575Z (string) (len=20) "aggregator-client-ca", 2025-12-04T11:45:12.196832575Z (string) (len=9) "client-ca" 2025-12-04T11:45:12.196832575Z }, 2025-12-04T11:45:12.196832575Z OptionalCertConfigMapNamePrefixes: ([]string) (len=1 cap=1) { 2025-12-04T11:45:12.196832575Z (string) (len=17) "trusted-ca-bundle" 2025-12-04T11:45:12.196832575Z }, 2025-12-04T11:45:12.196832575Z CertDir: (string) (len=66) "/etc/kubernetes/static-pod-resources/kube-controller-manager-certs", 2025-12-04T11:45:12.196832575Z ResourceDir: (string) (len=36) "/etc/kubernetes/static-pod-resources", 2025-12-04T11:45:12.196832575Z PodManifestDir: (string) (len=25) "/etc/kubernetes/manifests", 2025-12-04T11:45:12.196832575Z Timeout: (time.Duration) 2m0s, 2025-12-04T11:45:12.196832575Z StaticPodManifestsLockFile: (string) "", 2025-12-04T11:45:12.196832575Z PodMutationFns: ([]installerpod.PodMutationFunc) , 2025-12-04T11:45:12.196832575Z KubeletVersion: (string) "" 2025-12-04T11:45:12.196832575Z }) 2025-12-04T11:45:12.198331053Z I1204 11:45:12.198279 1 cmd.go:413] Getting controller reference for node master-0 2025-12-04T11:45:12.208893358Z I1204 11:45:12.207460 1 cmd.go:426] Waiting for installer revisions to settle for node master-0 2025-12-04T11:45:12.208893358Z I1204 11:45:12.207537 1 envvar.go:172] "Feature gate default state" feature="WatchListClient" enabled=false 2025-12-04T11:45:12.208893358Z I1204 11:45:12.207551 1 envvar.go:172] "Feature gate default state" feature="InformerResourceVersion" enabled=false 2025-12-04T11:45:12.213893022Z I1204 11:45:12.213824 1 cmd.go:518] Waiting additional period after revisions have settled for node master-0 2025-12-04T11:45:42.214130756Z I1204 11:45:42.214034 1 cmd.go:524] Getting installer pods for node master-0 2025-12-04T11:45:42.218346973Z I1204 11:45:42.218261 1 cmd.go:542] Latest installer revision for node master-0 is: 2 2025-12-04T11:45:42.218346973Z I1204 11:45:42.218301 1 cmd.go:431] Querying kubelet version for node master-0 2025-12-04T11:45:42.221984662Z I1204 11:45:42.221876 1 cmd.go:444] Got kubelet version 1.31.13 on target node master-0 2025-12-04T11:45:42.221984662Z I1204 11:45:42.221930 1 cmd.go:293] Creating target resource directory "/etc/kubernetes/static-pod-resources/kube-controller-manager-pod-2" ... 2025-12-04T11:45:42.222093606Z I1204 11:45:42.222064 1 cmd.go:221] Creating target resource directory "/etc/kubernetes/static-pod-resources/kube-controller-manager-pod-2" ... 2025-12-04T11:45:42.222093606Z I1204 11:45:42.222076 1 cmd.go:229] Getting secrets ... 2025-12-04T11:45:42.225434394Z I1204 11:45:42.225394 1 copy.go:32] Got secret openshift-kube-controller-manager/localhost-recovery-client-token-2 2025-12-04T11:45:42.227639087Z I1204 11:45:42.227591 1 copy.go:32] Got secret openshift-kube-controller-manager/service-account-private-key-2 2025-12-04T11:45:42.229789337Z I1204 11:45:42.229746 1 copy.go:32] Got secret openshift-kube-controller-manager/serving-cert-2 2025-12-04T11:45:42.229841468Z I1204 11:45:42.229793 1 cmd.go:242] Getting config maps ... 2025-12-04T11:45:42.233152917Z I1204 11:45:42.233112 1 copy.go:60] Got configMap openshift-kube-controller-manager/cluster-policy-controller-config-2 2025-12-04T11:45:42.235158542Z I1204 11:45:42.235133 1 copy.go:60] Got configMap openshift-kube-controller-manager/config-2 2025-12-04T11:45:42.238447759Z I1204 11:45:42.238215 1 copy.go:60] Got configMap openshift-kube-controller-manager/controller-manager-kubeconfig-2 2025-12-04T11:45:42.241775058Z I1204 11:45:42.241715 1 copy.go:60] Got configMap openshift-kube-controller-manager/kube-controller-cert-syncer-kubeconfig-2 2025-12-04T11:45:42.299597806Z I1204 11:45:42.299513 1 copy.go:60] Got configMap openshift-kube-controller-manager/kube-controller-manager-pod-2 2025-12-04T11:45:42.418973293Z I1204 11:45:42.418864 1 copy.go:60] Got configMap openshift-kube-controller-manager/recycler-config-2 2025-12-04T11:45:42.619926105Z I1204 11:45:42.619835 1 copy.go:60] Got configMap openshift-kube-controller-manager/service-ca-2 2025-12-04T11:45:42.818632573Z I1204 11:45:42.818540 1 copy.go:60] Got configMap openshift-kube-controller-manager/serviceaccount-ca-2 2025-12-04T11:45:43.019549932Z I1204 11:45:43.019438 1 copy.go:52] Failed to get config map openshift-kube-controller-manager/cloud-config-2: configmaps "cloud-config-2" not found 2025-12-04T11:45:43.019549932Z I1204 11:45:43.019492 1 cmd.go:261] Creating directory "/etc/kubernetes/static-pod-resources/kube-controller-manager-pod-2/secrets/localhost-recovery-client-token" ... 2025-12-04T11:45:43.019772090Z I1204 11:45:43.019734 1 cmd.go:639] Writing secret manifest "/etc/kubernetes/static-pod-resources/kube-controller-manager-pod-2/secrets/localhost-recovery-client-token/ca.crt" ... 2025-12-04T11:45:43.020034848Z I1204 11:45:43.019985 1 cmd.go:639] Writing secret manifest "/etc/kubernetes/static-pod-resources/kube-controller-manager-pod-2/secrets/localhost-recovery-client-token/namespace" ... 2025-12-04T11:45:43.020140822Z I1204 11:45:43.020102 1 cmd.go:639] Writing secret manifest "/etc/kubernetes/static-pod-resources/kube-controller-manager-pod-2/secrets/localhost-recovery-client-token/token" ... 2025-12-04T11:45:43.020248735Z I1204 11:45:43.020213 1 cmd.go:261] Creating directory "/etc/kubernetes/static-pod-resources/kube-controller-manager-pod-2/secrets/service-account-private-key" ... 2025-12-04T11:45:43.020314347Z I1204 11:45:43.020281 1 cmd.go:639] Writing secret manifest "/etc/kubernetes/static-pod-resources/kube-controller-manager-pod-2/secrets/service-account-private-key/service-account.key" ... 2025-12-04T11:45:43.020430831Z I1204 11:45:43.020398 1 cmd.go:261] Creating directory "/etc/kubernetes/static-pod-resources/kube-controller-manager-pod-2/secrets/serving-cert" ... 2025-12-04T11:45:43.020498014Z I1204 11:45:43.020463 1 cmd.go:639] Writing secret manifest "/etc/kubernetes/static-pod-resources/kube-controller-manager-pod-2/secrets/serving-cert/tls.crt" ... 2025-12-04T11:45:43.020622837Z I1204 11:45:43.020589 1 cmd.go:639] Writing secret manifest "/etc/kubernetes/static-pod-resources/kube-controller-manager-pod-2/secrets/serving-cert/tls.key" ... 2025-12-04T11:45:43.020710060Z I1204 11:45:43.020679 1 cmd.go:277] Creating directory "/etc/kubernetes/static-pod-resources/kube-controller-manager-pod-2/configmaps/cluster-policy-controller-config" ... 2025-12-04T11:45:43.020880355Z I1204 11:45:43.020799 1 cmd.go:629] Writing config file "/etc/kubernetes/static-pod-resources/kube-controller-manager-pod-2/configmaps/cluster-policy-controller-config/config.yaml" ... 2025-12-04T11:45:43.021005369Z I1204 11:45:43.020984 1 cmd.go:277] Creating directory "/etc/kubernetes/static-pod-resources/kube-controller-manager-pod-2/configmaps/config" ... 2025-12-04T11:45:43.021077772Z I1204 11:45:43.021051 1 cmd.go:629] Writing config file "/etc/kubernetes/static-pod-resources/kube-controller-manager-pod-2/configmaps/config/config.yaml" ... 2025-12-04T11:45:43.021186375Z I1204 11:45:43.021162 1 cmd.go:277] Creating directory "/etc/kubernetes/static-pod-resources/kube-controller-manager-pod-2/configmaps/controller-manager-kubeconfig" ... 2025-12-04T11:45:43.021239097Z I1204 11:45:43.021218 1 cmd.go:629] Writing config file "/etc/kubernetes/static-pod-resources/kube-controller-manager-pod-2/configmaps/controller-manager-kubeconfig/kubeconfig" ... 2025-12-04T11:45:43.021326820Z I1204 11:45:43.021304 1 cmd.go:277] Creating directory "/etc/kubernetes/static-pod-resources/kube-controller-manager-pod-2/configmaps/kube-controller-cert-syncer-kubeconfig" ... 2025-12-04T11:45:43.021414973Z I1204 11:45:43.021386 1 cmd.go:629] Writing config file "/etc/kubernetes/static-pod-resources/kube-controller-manager-pod-2/configmaps/kube-controller-cert-syncer-kubeconfig/kubeconfig" ... 2025-12-04T11:45:43.021539027Z I1204 11:45:43.021513 1 cmd.go:277] Creating directory "/etc/kubernetes/static-pod-resources/kube-controller-manager-pod-2/configmaps/kube-controller-manager-pod" ... 2025-12-04T11:45:43.021615719Z I1204 11:45:43.021593 1 cmd.go:629] Writing config file "/etc/kubernetes/static-pod-resources/kube-controller-manager-pod-2/configmaps/kube-controller-manager-pod/version" ... 2025-12-04T11:45:43.021726513Z I1204 11:45:43.021702 1 cmd.go:629] Writing config file "/etc/kubernetes/static-pod-resources/kube-controller-manager-pod-2/configmaps/kube-controller-manager-pod/forceRedeploymentReason" ... 2025-12-04T11:45:43.021863988Z I1204 11:45:43.021837 1 cmd.go:629] Writing config file "/etc/kubernetes/static-pod-resources/kube-controller-manager-pod-2/configmaps/kube-controller-manager-pod/pod.yaml" ... 2025-12-04T11:45:43.021987092Z I1204 11:45:43.021961 1 cmd.go:277] Creating directory "/etc/kubernetes/static-pod-resources/kube-controller-manager-pod-2/configmaps/recycler-config" ... 2025-12-04T11:45:43.022112376Z I1204 11:45:43.022077 1 cmd.go:629] Writing config file "/etc/kubernetes/static-pod-resources/kube-controller-manager-pod-2/configmaps/recycler-config/recycler-pod.yaml" ... 2025-12-04T11:45:43.022196219Z I1204 11:45:43.022173 1 cmd.go:277] Creating directory "/etc/kubernetes/static-pod-resources/kube-controller-manager-pod-2/configmaps/service-ca" ... 2025-12-04T11:45:43.022267581Z I1204 11:45:43.022245 1 cmd.go:629] Writing config file "/etc/kubernetes/static-pod-resources/kube-controller-manager-pod-2/configmaps/service-ca/ca-bundle.crt" ... 2025-12-04T11:45:43.022360774Z I1204 11:45:43.022339 1 cmd.go:277] Creating directory "/etc/kubernetes/static-pod-resources/kube-controller-manager-pod-2/configmaps/serviceaccount-ca" ... 2025-12-04T11:45:43.022415676Z I1204 11:45:43.022396 1 cmd.go:629] Writing config file "/etc/kubernetes/static-pod-resources/kube-controller-manager-pod-2/configmaps/serviceaccount-ca/ca-bundle.crt" ... 2025-12-04T11:45:43.022489128Z I1204 11:45:43.022468 1 cmd.go:221] Creating target resource directory "/etc/kubernetes/static-pod-resources/kube-controller-manager-certs" ... 2025-12-04T11:45:43.022538730Z I1204 11:45:43.022520 1 cmd.go:229] Getting secrets ... 2025-12-04T11:45:43.219568112Z I1204 11:45:43.219451 1 copy.go:32] Got secret openshift-kube-controller-manager/csr-signer 2025-12-04T11:45:43.419679046Z I1204 11:45:43.419580 1 copy.go:32] Got secret openshift-kube-controller-manager/kube-controller-manager-client-cert-key 2025-12-04T11:45:43.419679046Z I1204 11:45:43.419626 1 cmd.go:242] Getting config maps ... 2025-12-04T11:45:43.618301671Z I1204 11:45:43.618187 1 copy.go:60] Got configMap openshift-kube-controller-manager/aggregator-client-ca 2025-12-04T11:45:43.819588263Z I1204 11:45:43.818796 1 copy.go:60] Got configMap openshift-kube-controller-manager/client-ca 2025-12-04T11:45:44.026079165Z I1204 11:45:44.026017 1 copy.go:60] Got configMap openshift-kube-controller-manager/trusted-ca-bundle 2025-12-04T11:45:44.026355824Z I1204 11:45:44.026336 1 cmd.go:261] Creating directory "/etc/kubernetes/static-pod-resources/kube-controller-manager-certs/secrets/csr-signer" ... 2025-12-04T11:45:44.026515719Z I1204 11:45:44.026498 1 cmd.go:639] Writing secret manifest "/etc/kubernetes/static-pod-resources/kube-controller-manager-certs/secrets/csr-signer/tls.crt" ... 2025-12-04T11:45:44.026637113Z I1204 11:45:44.026624 1 cmd.go:639] Writing secret manifest "/etc/kubernetes/static-pod-resources/kube-controller-manager-certs/secrets/csr-signer/tls.key" ... 2025-12-04T11:45:44.026712886Z I1204 11:45:44.026700 1 cmd.go:261] Creating directory "/etc/kubernetes/static-pod-resources/kube-controller-manager-certs/secrets/kube-controller-manager-client-cert-key" ... 2025-12-04T11:45:44.026758537Z I1204 11:45:44.026748 1 cmd.go:639] Writing secret manifest "/etc/kubernetes/static-pod-resources/kube-controller-manager-certs/secrets/kube-controller-manager-client-cert-key/tls.key" ... 2025-12-04T11:45:44.026864071Z I1204 11:45:44.026850 1 cmd.go:639] Writing secret manifest "/etc/kubernetes/static-pod-resources/kube-controller-manager-certs/secrets/kube-controller-manager-client-cert-key/tls.crt" ... 2025-12-04T11:45:44.026939783Z I1204 11:45:44.026928 1 cmd.go:277] Creating directory "/etc/kubernetes/static-pod-resources/kube-controller-manager-certs/configmaps/aggregator-client-ca" ... 2025-12-04T11:45:44.027024746Z I1204 11:45:44.027012 1 cmd.go:629] Writing config file "/etc/kubernetes/static-pod-resources/kube-controller-manager-certs/configmaps/aggregator-client-ca/ca-bundle.crt" ... 2025-12-04T11:45:44.027101579Z I1204 11:45:44.027090 1 cmd.go:277] Creating directory "/etc/kubernetes/static-pod-resources/kube-controller-manager-certs/configmaps/client-ca" ... 2025-12-04T11:45:44.027149870Z I1204 11:45:44.027138 1 cmd.go:629] Writing config file "/etc/kubernetes/static-pod-resources/kube-controller-manager-certs/configmaps/client-ca/ca-bundle.crt" ... 2025-12-04T11:45:44.027235833Z I1204 11:45:44.027223 1 cmd.go:277] Creating directory "/etc/kubernetes/static-pod-resources/kube-controller-manager-certs/configmaps/trusted-ca-bundle" ... 2025-12-04T11:45:44.027322116Z I1204 11:45:44.027310 1 cmd.go:629] Writing config file "/etc/kubernetes/static-pod-resources/kube-controller-manager-certs/configmaps/trusted-ca-bundle/ca-bundle.crt" ... 2025-12-04T11:45:44.028155113Z I1204 11:45:44.028124 1 cmd.go:335] Getting pod configmaps/kube-controller-manager-pod-2 -n openshift-kube-controller-manager 2025-12-04T11:45:44.218346263Z I1204 11:45:44.218270 1 cmd.go:351] Creating directory for static pod manifest "/etc/kubernetes/manifests" ... 2025-12-04T11:45:44.218408445Z I1204 11:45:44.218332 1 cmd.go:379] Writing a pod under "kube-controller-manager-pod.yaml" key 2025-12-04T11:45:44.218408445Z {"kind":"Pod","apiVersion":"v1","metadata":{"name":"kube-controller-manager","namespace":"openshift-kube-controller-manager","creationTimestamp":null,"labels":{"app":"kube-controller-manager","kube-controller-manager":"true","revision":"2"},"annotations":{"kubectl.kubernetes.io/default-container":"kube-controller-manager","target.workload.openshift.io/management":"{\"effect\": \"PreferredDuringScheduling\"}"}},"spec":{"volumes":[{"name":"resource-dir","hostPath":{"path":"/etc/kubernetes/static-pod-resources/kube-controller-manager-pod-2"}},{"name":"cert-dir","hostPath":{"path":"/etc/kubernetes/static-pod-resources/kube-controller-manager-certs"}}],"containers":[{"name":"kube-controller-manager","image":"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:6d5891cdd7dcf7c9081de8b364b4c96446b7f946f7880fbae291a4592a198264","command":["/bin/bash","-euxo","pipefail","-c"],"args":["timeout 3m /bin/bash -exuo pipefail -c 'while [ -n \"$(ss -Htanop \\( sport = 10257 \\))\" ]; do sleep 1; done'\n\nif [ -f /etc/kubernetes/static-pod-certs/configmaps/trusted-ca-bundle/ca-bundle.crt ]; then\n echo \"Copying system trust bundle\"\n cp -f /etc/kubernetes/static-pod-certs/configmaps/trusted-ca-bundle/ca-bundle.crt /etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem\nfi\n\nif [ -f /etc/kubernetes/static-pod-resources/configmaps/cloud-config/ca-bundle.pem ]; then\n echo \"Setting custom CA bundle for cloud provider\"\n export AWS_CA_BUNDLE=/etc/kubernetes/static-pod-resources/configmaps/cloud-config/ca-bundle.pem\nfi\n\nexec hyperkube kube-controller-manager --openshift-config=/etc/kubernetes/static-pod-resources/configmaps/config/config.yaml \\\n --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig \\\n --authentication-kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig \\\n --authorization-kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig \\\n --client-ca-file=/etc/kubernetes/static-pod-certs/configmaps/client-ca/ca-bundle.crt \\\n --requestheader-client-ca-file=/etc/kubernetes/static-pod-certs/configmaps/aggregator-client-ca/ca-bundle.crt -v=2 --allocate-node-cidrs=false --cert-dir=/var/run/kubernetes --cloud-provider=external --cluster-cidr=10.128.0.0/16 --cluster-name=sno-d7v9r --cluster-signing-cert-file=/etc/kubernetes/static-pod-certs/secrets/csr-signer/tls.crt --cluster-signing-duration=720h --cluster-signing-key-file=/etc/kubernetes/static-pod-certs/secrets/csr-signer/tls.key --controllers=* --controllers=-bootstrapsigner --controllers=-tokencleaner --controllers=-ttl --enable-dynamic-provisioning=true --feature-gates=AWSClusterHostedDNS=false --feature-gates=AWSEFSDriverVolumeMetrics=true --feature-gates=AdditionalRoutingCapabilities=false --feature-gates=AdminNetworkPolicy=true --feature-gates=AlibabaPlatform=true --feature-gates=AutomatedEtcdBackup=false --feature-gates=AzureWorkloadIdentity=true --feature-gates=BareMetalLoadBalancer=true --feature-gates=BootcNodeManagement=false --feature-gates=BuildCSIVolumes=true --feature-gates=CSIDriverSharedResource=false --feature-gates=ChunkSizeMiB=true --feature-gates=CloudDualStackNodeIPs=true --feature-gates=ClusterAPIInstall=false --feature-gates=ClusterAPIInstallIBMCloud=false --feature-gates=ClusterMonitoringConfig=false --feature-gates=ConsolePluginContentSecurityPolicy=false --feature-gates=DNSNameResolver=false --feature-gates=DisableKubeletCloudCredentialProviders=true --feature-gates=DynamicResourceAllocation=false --feature-gates=EtcdBackendQuota=false --feature-gates=EventedPLEG=false --feature-gates=Example=false --feature-gates=ExternalOIDC=false --feature-gates=ExternalOIDCWithUIDAndExtraClaimMappings=false --feature-gates=GCPClusterHostedDNS=false --feature-gates=GCPLabelsTags=true --feature-gates=GatewayAPI=false --feature-gates=HardwareSpeed=true --feature-gates=ImageStreamImportMode=false --feature-gates=IngressControllerDynamicConfigurationManager=false --feature-gates=IngressControllerLBSubnetsAWS=true --feature-gates=InsightsConfig=false --feature-gates=InsightsConfigAPI=false --feature-gates=InsightsOnDemandDataGather=false --feature-gates=InsightsRuntimeExtractor=false --feature-gates=KMSv1=true --feature-gates=MachineAPIMigration=false --feature-gates=MachineAPIOperatorDisableMachineHealthCheckController=false --feature-gates=MachineAPIProviderOpenStack=false --feature-gates=MachineConfigNodes=false --feature-gates=ManagedBootImages=true --feature-gates=ManagedBootImagesAWS=true --feature-gates=MaxUnavailableStatefulSet=false --feature-gates=MetricsCollectionProfiles=false --feature-gates=MinimumKubeletVersion=false --feature-gates=MixedCPUsAllocation=false --feature-gates=MultiArchInstallAWS=true --feature-gates=MultiArchInstallAzure=false --feature-gates=MultiArchInstallGCP=true --feature-gates=NetworkDiagnosticsConfig=true --feature-gates=NetworkLiveMigration=true --feature-gates=NetworkSegmentation=true --feature-gates=NewOLM=true --feature-gates=NodeDisruptionPolicy=true --feature-gates=NodeSwap=false --feature-gates=NutanixMultiSubnets=false --feature-gates=OVNObservability=false --feature-gates=OnClusterBuild=true --feature-gates=OpenShiftPodSecurityAdmission=false --feature-gates=PersistentIPsForVirtualization=true --feature-gates=PinnedImages=false --feature-gates=PlatformOperators=false --feature-gates=PrivateHostedZoneAWS=true --feature-gates=ProcMountType=false --feature-gates=RouteAdvertisements=false --feature-gates=RouteExternalCertificate=false --feature-gates=ServiceAccountTokenNodeBinding=false --feature-gates=SetEIPForNLBIngressController=true --feature-gates=SignatureStores=false --feature-gates=SigstoreImageVerification=false --feature-gates=StreamingCollectionEncodingToJSON=false --feature-gates=StreamingCollectionEncodingToProtobuf=false --feature-gates=TranslateStreamCloseWebsocketRequests=false --feature-gates=UpgradeStatus=false --feature-gates=UserNamespacesPodSecurityStandards=false --feature-gates=UserNamespacesSupport=false --feature-gates=VSphereControlPlaneMachineSet=true --feature-gates=VSphereDriverConfiguration=true --feature-gates=VSphereMultiNetworks=false --feature-gates=VSphereMultiVCenters=true --feature-gates=VSphereStaticIPs=true --feature-gates=ValidatingAdmissionPolicy=true --feature-gates=VolumeAttributesClass=false --feature-gates=VolumeGroupSnapshot=false --flex-volume-plugin-dir=/etc/kubernetes/kubelet-plugins/volume/exec --kube-api-burst=300 --kube-api-qps=150 --leader-elect-renew-deadline=12s --leader-elect-resource-lock=leases --leader-elect-retry-period=3s --leader-elect=true --pv-recycler-pod-template-filepath-hostpath=/etc/kubernetes/static-pod-resources/configmaps/recycler-config/recycler-pod.yaml --pv-recycler-pod-template-filepath-nfs=/etc/kubernetes/static-pod-resources/configmaps/recycler-config/recycler-pod.yaml --root-ca-file=/etc/kubernetes/static-pod-resources/configmaps/serviceaccount-ca/ca-bundle.crt --secure-port=10257 --service-account-private-key-file=/etc/kubernetes/static-pod-resources/secrets/service-account-private-key/service-account.key --service-cluster-ip-range=172.30.0.0/16 --use-service-account-credentials=true --tls-cipher-suites=TLS_AES_128_GCM_SHA256,TLS_AES_256_GCM_SHA384,TLS_CHACHA20_POLY1305_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305_SHA256,TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305_SHA256 --tls-min-version=VersionTLS12"],"ports":[{"containerPort":10257}],"resources":{"requests":{"cpu":"60m","memory":"200Mi"}},"volumeMounts":[{"name":"resource-dir","mountPath":"/etc/kubernetes/static-pod-resources"},{"name":"cert-dir","mountPath":"/etc/kubernetes/static-pod-certs"}],"livenessProbe":{"httpGet":{"path":"healthz","port":10257,"scheme":"HTTPS"},"initialDelaySeconds":45,"timeoutSeconds":10},"readinessProbe":{"httpGet":{"path":"healthz","port":10257,"scheme":"HTTPS"},"initialDelaySeconds":10,"timeoutSeconds":10},"startupProbe":{"httpGet":{"path":"healthz","port":10257,"scheme":"HTTPS"},"timeoutSeconds":3},"terminationMessagePolicy":"FallbackToLogsOnError","imagePullPolicy":"IfNotPresent"},{"name":"cluster-policy-controller","image":"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d64c13fe7663a0b4ae61d103b1b7598adcf317a01826f296bcb66b1a2de83c96","command":["/bin/bash","-euxo","pipefail","-c"],"args":["timeout 3m /bin/bash -exuo pipefail -c 'while [ -n \"$(ss -Htanop \\( sport = 10357 \\))\" ]; do sleep 1; done'\n\nexec cluster-policy-controller start --config=/etc/kubernetes/static-pod-resources/configmaps/cluster-policy-controller-config/config.yaml \\\n --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig \\\n --namespace=${POD_NAMESPACE} -v=2"],"ports":[{"containerPort":10357}],"env":[{"name":"POD_NAME","valueFrom":{"fieldRef":{"fieldPath":"metadata.name"}}},{"name":"POD_NAMESPACE","valueFrom":{"fieldRef":{"fieldPath":"metadata.namespace"}}}],"resources":{"requests":{"cpu":"10m","memory":"200Mi"}},"volumeMounts":[{"name":"resource-dir","mountPath":"/etc/kubernetes/static-pod-resources"},{"name":"cert-dir","mountPath":"/etc/kubernetes/static-pod-certs"}],"livenessProbe":{"httpGet":{"path":"healthz","port":10357,"host":"localhost","scheme":"HTTPS"},"initialDelaySeconds":45,"timeoutSeconds":10},"readinessProbe":{"httpGet":{"path":"healthz","port":10357,"host":"localhost","scheme":"HTTPS"},"initialDelaySeconds":10,"timeoutSeconds":10},"startupProbe":{"httpGet":{"path":"healthz","port":10357,"host":"localhost","scheme":"HTTPS"},"timeoutSeconds":3},"terminationMessagePolicy":"FallbackToLogsOnError","imagePullPolicy":"IfNotPresent"},{"name":"kube-controller-manager-cert-syncer","image":"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a824e468cf8dd61d347e35b2ee5bc2f815666957647098e21a1bb56ff613e5b9","command":["cluster-kube-controller-manager-operator","cert-syncer"],"args":["--kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/kube-controller-cert-syncer-kubeconfig/kubeconfig","--namespace=$(POD_NAMESPACE)","--destination-dir=/etc/kubernetes/static-pod-certs"],"env":[{"name":"POD_NAME","valueFrom":{"fieldRef":{"fieldPath":"metadata.name"}}},{"name":"POD_NAMESPACE","valueFrom":{"fieldRef":{"fieldPath":"metadata.namespace"}}}],"resources":{"requests":{"cpu":"5m","memory":"50Mi"}},"volumeMounts":[{"name":"resource-dir","mountPath":"/etc/kubernetes/static-pod-resources"},{"name":"cert-dir","mountPath":"/etc/kubernetes/static-pod-certs"}],"terminationMessagePolicy":"FallbackToLogsOnError","imagePullPolicy":"IfNotPresent"},{"name":"kube-controller-manager-recovery-controller","image":"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a824e468cf8dd61d347e35b2ee5bc2f815666957647098e21a1bb56ff613e5b9","command":["/bin/bash","-euxo","pipefail","-c"],"args":["timeout 3m /bin/bash -exuo pipefail -c 'while [ -n \"$(ss -Htanop \\( sport = 9443 \\))\" ]; do sleep 1; done'\n\nexec cluster-kube-controller-manager-operator cert-recovery-controller --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/kube-controller-cert-syncer-kubeconfig/kubeconfig --namespace=${POD_NAMESPACE} --listen=0.0.0.0:9443 -v=2"],"env":[{"name":"POD_NAMESPACE","valueFrom":{"fieldRef":{"fieldPath":"metadata.namespace"}}}],"resources":{"requests":{"cpu":"5m","memory":"50Mi"}},"volumeMounts":[{"name":"resource-dir","mountPath":"/etc/kubernetes/static-pod-resources"},{"name":"cert-dir","mountPath":"/etc/kubernetes/static-pod-certs"}],"terminationMessagePolicy":"FallbackToLogsOnError","imagePullPolicy":"IfNotPresent"}],"hostNetwork":true,"tolerations":[{"operator":"Exists"}],"priorityClassName":"system-node-critical"},"status":{}} 2025-12-04T11:45:44.225177666Z I1204 11:45:44.225108 1 cmd.go:610] Writing pod manifest "/etc/kubernetes/static-pod-resources/kube-controller-manager-pod-2/kube-controller-manager-pod.yaml" ... 2025-12-04T11:45:44.226134707Z I1204 11:45:44.226113 1 cmd.go:617] Removed existing static pod manifest "/etc/kubernetes/manifests/kube-controller-manager-pod.yaml" ... 2025-12-04T11:45:44.226195649Z I1204 11:45:44.226166 1 cmd.go:621] Writing static pod manifest "/etc/kubernetes/manifests/kube-controller-manager-pod.yaml" ... 2025-12-04T11:45:44.226195649Z {"kind":"Pod","apiVersion":"v1","metadata":{"name":"kube-controller-manager","namespace":"openshift-kube-controller-manager","creationTimestamp":null,"labels":{"app":"kube-controller-manager","kube-controller-manager":"true","revision":"2"},"annotations":{"kubectl.kubernetes.io/default-container":"kube-controller-manager","target.workload.openshift.io/management":"{\"effect\": \"PreferredDuringScheduling\"}"}},"spec":{"volumes":[{"name":"resource-dir","hostPath":{"path":"/etc/kubernetes/static-pod-resources/kube-controller-manager-pod-2"}},{"name":"cert-dir","hostPath":{"path":"/etc/kubernetes/static-pod-resources/kube-controller-manager-certs"}}],"containers":[{"name":"kube-controller-manager","image":"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:6d5891cdd7dcf7c9081de8b364b4c96446b7f946f7880fbae291a4592a198264","command":["/bin/bash","-euxo","pipefail","-c"],"args":["timeout 3m /bin/bash -exuo pipefail -c 'while [ -n \"$(ss -Htanop \\( sport = 10257 \\))\" ]; do sleep 1; done'\n\nif [ -f /etc/kubernetes/static-pod-certs/configmaps/trusted-ca-bundle/ca-bundle.crt ]; then\n echo \"Copying system trust bundle\"\n cp -f /etc/kubernetes/static-pod-certs/configmaps/trusted-ca-bundle/ca-bundle.crt /etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem\nfi\n\nif [ -f /etc/kubernetes/static-pod-resources/configmaps/cloud-config/ca-bundle.pem ]; then\n echo \"Setting custom CA bundle for cloud provider\"\n export AWS_CA_BUNDLE=/etc/kubernetes/static-pod-resources/configmaps/cloud-config/ca-bundle.pem\nfi\n\nexec hyperkube kube-controller-manager --openshift-config=/etc/kubernetes/static-pod-resources/configmaps/config/config.yaml \\\n --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig \\\n --authentication-kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig \\\n --authorization-kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig \\\n --client-ca-file=/etc/kubernetes/static-pod-certs/configmaps/client-ca/ca-bundle.crt \\\n --requestheader-client-ca-file=/etc/kubernetes/static-pod-certs/configmaps/aggregator-client-ca/ca-bundle.crt -v=2 --allocate-node-cidrs=false --cert-dir=/var/run/kubernetes --cloud-provider=external --cluster-cidr=10.128.0.0/16 --cluster-name=sno-d7v9r --cluster-signing-cert-file=/etc/kubernetes/static-pod-certs/secrets/csr-signer/tls.crt --cluster-signing-duration=720h --cluster-signing-key-file=/etc/kubernetes/static-pod-certs/secrets/csr-signer/tls.key --controllers=* --controllers=-bootstrapsigner --controllers=-tokencleaner --controllers=-ttl --enable-dynamic-provisioning=true --feature-gates=AWSClusterHostedDNS=false --feature-gates=AWSEFSDriverVolumeMetrics=true --feature-gates=AdditionalRoutingCapabilities=false --feature-gates=AdminNetworkPolicy=true --feature-gates=AlibabaPlatform=true --feature-gates=AutomatedEtcdBackup=false --feature-gates=AzureWorkloadIdentity=true --feature-gates=BareMetalLoadBalancer=true --feature-gates=BootcNodeManagement=false --feature-gates=BuildCSIVolumes=true --feature-gates=CSIDriverSharedResource=false --feature-gates=ChunkSizeMiB=true --feature-gates=CloudDualStackNodeIPs=true --feature-gates=ClusterAPIInstall=false --feature-gates=ClusterAPIInstallIBMCloud=false --feature-gates=ClusterMonitoringConfig=false --feature-gates=ConsolePluginContentSecurityPolicy=false --feature-gates=DNSNameResolver=false --feature-gates=DisableKubeletCloudCredentialProviders=true --feature-gates=DynamicResourceAllocation=false --feature-gates=EtcdBackendQuota=false --feature-gates=EventedPLEG=false --feature-gates=Example=false --feature-gates=ExternalOIDC=false --feature-gates=ExternalOIDCWithUIDAndExtraClaimMappings=false --feature-gates=GCPClusterHostedDNS=false --feature-gates=GCPLabelsTags=true --feature-gates=GatewayAPI=false --feature-gates=HardwareSpeed=true --feature-gates=ImageStreamImportMode=false --feature-gates=IngressControllerDynamicConfigurationManager=false --feature-gates=IngressControllerLBSubnetsAWS=true --feature-gates=InsightsConfig=false --feature-gates=InsightsConfigAPI=false --feature-gates=InsightsOnDemandDataGather=false --feature-gates=InsightsRuntimeExtractor=false --feature-gates=KMSv1=true --feature-gates=MachineAPIMigration=false --feature-gates=MachineAPIOperatorDisableMachineHealthCheckController=false --feature-gates=MachineAPIProviderOpenStack=false --feature-gates=MachineConfigNodes=false --feature-gates=ManagedBootImages=true --feature-gates=ManagedBootImagesAWS=true --feature-gates=MaxUnavailableStatefulSet=false --feature-gates=MetricsCollectionProfiles=false --feature-gates=MinimumKubeletVersion=false --feature-gates=MixedCPUsAllocation=false --feature-gates=MultiArchInstallAWS=true --feature-gates=MultiArchInstallAzure=false --feature-gates=MultiArchInstallGCP=true --feature-gates=NetworkDiagnosticsConfig=true --feature-gates=NetworkLiveMigration=true --feature-gates=NetworkSegmentation=true --feature-gates=NewOLM=true --feature-gates=NodeDisruptionPolicy=true --feature-gates=NodeSwap=false --feature-gates=NutanixMultiSubnets=false --feature-gates=OVNObservability=false --feature-gates=OnClusterBuild=true --feature-gates=OpenShiftPodSecurityAdmission=false --feature-gates=PersistentIPsForVirtualization=true --feature-gates=PinnedImages=false --feature-gates=PlatformOperators=false --feature-gates=PrivateHostedZoneAWS=true --feature-gates=ProcMountType=false --feature-gates=RouteAdvertisements=false --feature-gates=RouteExternalCertificate=false --feature-gates=ServiceAccountTokenNodeBinding=false --feature-gates=SetEIPForNLBIngressController=true --feature-gates=SignatureStores=false --feature-gates=SigstoreImageVerification=false --feature-gates=StreamingCollectionEncodingToJSON=false --feature-gates=StreamingCollectionEncodingToProtobuf=false --feature-gates=TranslateStreamCloseWebsocketRequests=false --feature-gates=UpgradeStatus=false --feature-gates=UserNamespacesPodSecurityStandards=false --feature-gates=UserNamespacesSupport=false --feature-gates=VSphereControlPlaneMachineSet=true --feature-gates=VSphereDriverConfiguration=true --feature-gates=VSphereMultiNetworks=false --feature-gates=VSphereMultiVCenters=true --feature-gates=VSphereStaticIPs=true --feature-gates=ValidatingAdmissionPolicy=true --feature-gates=VolumeAttributesClass=false --feature-gates=VolumeGroupSnapshot=false --flex-volume-plugin-dir=/etc/kubernetes/kubelet-plugins/volume/exec --kube-api-burst=300 --kube-api-qps=150 --leader-elect-renew-deadline=12s --leader-elect-resource-lock=leases --leader-elect-retry-period=3s --leader-elect=true --pv-recycler-pod-template-filepath-hostpath=/etc/kubernetes/static-pod-resources/configmaps/recycler-config/recycler-pod.yaml --pv-recycler-pod-template-filepath-nfs=/etc/kubernetes/static-pod-resources/configmaps/recycler-config/recycler-pod.yaml --root-ca-file=/etc/kubernetes/static-pod-resources/configmaps/serviceaccount-ca/ca-bundle.crt --secure-port=10257 --service-account-private-key-file=/etc/kubernetes/static-pod-resources/secrets/service-account-private-key/service-account.key --service-cluster-ip-range=172.30.0.0/16 --use-service-account-credentials=true --tls-cipher-suites=TLS_AES_128_GCM_SHA256,TLS_AES_256_GCM_SHA384,TLS_CHACHA20_POLY1305_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305_SHA256,TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305_SHA256 --tls-min-version=VersionTLS12"],"ports":[{"containerPort":10257}],"resources":{"requests":{"cpu":"60m","memory":"200Mi"}},"volumeMounts":[{"name":"resource-dir","mountPath":"/etc/kubernetes/static-pod-resources"},{"name":"cert-dir","mountPath":"/etc/kubernetes/static-pod-certs"}],"livenessProbe":{"httpGet":{"path":"healthz","port":10257,"scheme":"HTTPS"},"initialDelaySeconds":45,"timeoutSeconds":10},"readinessProbe":{"httpGet":{"path":"healthz","port":10257,"scheme":"HTTPS"},"initialDelaySeconds":10,"timeoutSeconds":10},"startupProbe":{"httpGet":{"path":"healthz","port":10257,"scheme":"HTTPS"},"timeoutSeconds":3},"terminationMessagePolicy":"FallbackToLogsOnError","imagePullPolicy":"IfNotPresent"},{"name":"cluster-policy-controller","image":"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d64c13fe7663a0b4ae61d103b1b7598adcf317a01826f296bcb66b1a2de83c96","command":["/bin/bash","-euxo","pipefail","-c"],"args":["timeout 3m /bin/bash -exuo pipefail -c 'while [ -n \"$(ss -Htanop \\( sport = 10357 \\))\" ]; do sleep 1; done'\n\nexec cluster-policy-controller start --config=/etc/kubernetes/static-pod-resources/configmaps/cluster-policy-controller-config/config.yaml \\\n --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig \\\n --namespace=${POD_NAMESPACE} -v=2"],"ports":[{"containerPort":10357}],"env":[{"name":"POD_NAME","valueFrom":{"fieldRef":{"fieldPath":"metadata.name"}}},{"name":"POD_NAMESPACE","valueFrom":{"fieldRef":{"fieldPath":"metadata.namespace"}}}],"resources":{"requests":{"cpu":"10m","memory":"200Mi"}},"volumeMounts":[{"name":"resource-dir","mountPath":"/etc/kubernetes/static-pod-resources"},{"name":"cert-dir","mountPath":"/etc/kubernetes/static-pod-certs"}],"livenessProbe":{"httpGet":{"path":"healthz","port":10357,"host":"localhost","scheme":"HTTPS"},"initialDelaySeconds":45,"timeoutSeconds":10},"readinessProbe":{"httpGet":{"path":"healthz","port":10357,"host":"localhost","scheme":"HTTPS"},"initialDelaySeconds":10,"timeoutSeconds":10},"startupProbe":{"httpGet":{"path":"healthz","port":10357,"host":"localhost","scheme":"HTTPS"},"timeoutSeconds":3},"terminationMessagePolicy":"FallbackToLogsOnError","imagePullPolicy":"IfNotPresent"},{"name":"kube-controller-manager-cert-syncer","image":"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a824e468cf8dd61d347e35b2ee5bc2f815666957647098e21a1bb56ff613e5b9","command":["cluster-kube-controller-manager-operator","cert-syncer"],"args":["--kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/kube-controller-cert-syncer-kubeconfig/kubeconfig","--namespace=$(POD_NAMESPACE)","--destination-dir=/etc/kubernetes/static-pod-certs"],"env":[{"name":"POD_NAME","valueFrom":{"fieldRef":{"fieldPath":"metadata.name"}}},{"name":"POD_NAMESPACE","valueFrom":{"fieldRef":{"fieldPath":"metadata.namespace"}}}],"resources":{"requests":{"cpu":"5m","memory":"50Mi"}},"volumeMounts":[{"name":"resource-dir","mountPath":"/etc/kubernetes/static-pod-resources"},{"name":"cert-dir","mountPath":"/etc/kubernetes/static-pod-certs"}],"terminationMessagePolicy":"FallbackToLogsOnError","imagePullPolicy":"IfNotPresent"},{"name":"kube-controller-manager-recovery-controller","image":"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a824e468cf8dd61d347e35b2ee5bc2f815666957647098e21a1bb56ff613e5b9","command":["/bin/bash","-euxo","pipefail","-c"],"args":["timeout 3m /bin/bash -exuo pipefail -c 'while [ -n \"$(ss -Htanop \\( sport = 9443 \\))\" ]; do sleep 1; done'\n\nexec cluster-kube-controller-manager-operator cert-recovery-controller --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/kube-controller-cert-syncer-kubeconfig/kubeconfig --namespace=${POD_NAMESPACE} --listen=0.0.0.0:9443 -v=2"],"env":[{"name":"POD_NAMESPACE","valueFrom":{"fieldRef":{"fieldPath":"metadata.namespace"}}}],"resources":{"requests":{"cpu":"5m","memory":"50Mi"}},"volumeMounts":[{"name":"resource-dir","mountPath":"/etc/kubernetes/static-pod-resources"},{"name":"cert-dir","mountPath":"/etc/kubernetes/static-pod-certs"}],"terminationMessagePolicy":"FallbackToLogsOnError","imagePullPolicy":"IfNotPresent"}],"hostNetwork":true,"tolerations":[{"operator":"Exists"}],"priorityClassName":"system-node-critical"},"status":{}} config/pod/openshift-ingress/router-default-5465c8b4db-58d52.json0000640000000000000000000001534115114274105022734 0ustar0000000000000000{"metadata":{"name":"router-default-5465c8b4db-58d52","generateName":"router-default-5465c8b4db-","namespace":"openshift-ingress","uid":"b321ddb3-08ce-4101-bfa3-4200006f11ac","resourceVersion":"6588","creationTimestamp":"2025-12-04T11:38:21Z","labels":{"ingresscontroller.operator.openshift.io/deployment-ingresscontroller":"default","ingresscontroller.operator.openshift.io/hash":"85fc95cf79","pod-template-hash":"5465c8b4db"},"annotations":{"openshift.io/required-scc":"hostnetwork","openshift.io/scc":"hostnetwork"},"ownerReferences":[{"apiVersion":"apps/v1","kind":"ReplicaSet","name":"router-default-5465c8b4db","uid":"a7b90f45-efb1-4b60-931e-e22cf8de6436","controller":true,"blockOwnerDeletion":true}]},"spec":{"volumes":[{"name":"default-certificate","secret":{"secretName":"router-certs-default","defaultMode":420}},{"name":"service-ca-bundle","configMap":{"name":"service-ca-bundle","items":[{"key":"service-ca.crt","path":"service-ca.crt"}],"defaultMode":420,"optional":false}},{"name":"stats-auth","secret":{"secretName":"router-stats-default","defaultMode":420}},{"name":"metrics-certs","secret":{"secretName":"router-metrics-certs-default","defaultMode":420}},{"name":"kube-api-access-4clj7","projected":{"sources":[{"serviceAccountToken":{"expirationSeconds":3607,"path":"token"}},{"configMap":{"name":"kube-root-ca.crt","items":[{"key":"ca.crt","path":"ca.crt"}]}},{"downwardAPI":{"items":[{"path":"namespace","fieldRef":{"apiVersion":"v1","fieldPath":"metadata.namespace"}}]}},{"configMap":{"name":"openshift-service-ca.crt","items":[{"key":"service-ca.crt","path":"service-ca.crt"}]}}],"defaultMode":420}}],"containers":[{"name":"router","image":"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:2b3d313c599852b3543ee5c3a62691bd2d1bbad12c2e1c610cd71a1dec6eea32","ports":[{"name":"http","hostPort":80,"containerPort":80,"protocol":"TCP"},{"name":"https","hostPort":443,"containerPort":443,"protocol":"TCP"},{"name":"metrics","hostPort":1936,"containerPort":1936,"protocol":"TCP"}],"env":[{"name":"DEFAULT_CERTIFICATE_DIR","value":"/etc/pki/tls/private"},{"name":"DEFAULT_DESTINATION_CA_PATH","value":"/var/run/configmaps/service-ca/service-ca.crt"},{"name":"RELOAD_INTERVAL","value":"5s"},{"name":"ROUTER_ALLOW_WILDCARD_ROUTES","value":"false"},{"name":"ROUTER_CANONICAL_HOSTNAME","value":"router-default.apps.sno.openstack.lab"},{"name":"ROUTER_CIPHERS","value":"ECDHE-ECDSA-AES128-GCM-SHA256:ECDHE-RSA-AES128-GCM-SHA256:ECDHE-ECDSA-AES256-GCM-SHA384:ECDHE-RSA-AES256-GCM-SHA384:ECDHE-ECDSA-CHACHA20-POLY1305:ECDHE-RSA-CHACHA20-POLY1305:DHE-RSA-AES128-GCM-SHA256:DHE-RSA-AES256-GCM-SHA384"},{"name":"ROUTER_CIPHERSUITES","value":"TLS_AES_128_GCM_SHA256:TLS_AES_256_GCM_SHA384:TLS_CHACHA20_POLY1305_SHA256"},{"name":"ROUTER_DISABLE_HTTP2","value":"true"},{"name":"ROUTER_DISABLE_NAMESPACE_OWNERSHIP_CHECK","value":"false"},{"name":"ROUTER_DOMAIN","value":"apps.sno.openstack.lab"},{"name":"ROUTER_IDLE_CLOSE_ON_RESPONSE","value":"true"},{"name":"ROUTER_LOAD_BALANCE_ALGORITHM","value":"random"},{"name":"ROUTER_METRICS_TLS_CERT_FILE","value":"/etc/pki/tls/metrics-certs/tls.crt"},{"name":"ROUTER_METRICS_TLS_KEY_FILE","value":"/etc/pki/tls/metrics-certs/tls.key"},{"name":"ROUTER_METRICS_TYPE","value":"haproxy"},{"name":"ROUTER_SERVICE_HTTPS_PORT","value":"443"},{"name":"ROUTER_SERVICE_HTTP_PORT","value":"80"},{"name":"ROUTER_SERVICE_NAME","value":"default"},{"name":"ROUTER_SERVICE_NAMESPACE","value":"openshift-ingress"},{"name":"ROUTER_SET_FORWARDED_HEADERS","value":"append"},{"name":"ROUTER_TCP_BALANCE_SCHEME","value":"source"},{"name":"ROUTER_THREADS","value":"4"},{"name":"SSL_MIN_VERSION","value":"TLSv1.2"},{"name":"STATS_PASSWORD_FILE","value":"/var/lib/haproxy/conf/metrics-auth/statsPassword"},{"name":"STATS_PORT","value":"1936"},{"name":"STATS_USERNAME_FILE","value":"/var/lib/haproxy/conf/metrics-auth/statsUsername"}],"resources":{"requests":{"cpu":"100m","memory":"256Mi"}},"volumeMounts":[{"name":"default-certificate","readOnly":true,"mountPath":"/etc/pki/tls/private"},{"name":"service-ca-bundle","readOnly":true,"mountPath":"/var/run/configmaps/service-ca"},{"name":"stats-auth","readOnly":true,"mountPath":"/var/lib/haproxy/conf/metrics-auth"},{"name":"metrics-certs","readOnly":true,"mountPath":"/etc/pki/tls/metrics-certs"},{"name":"kube-api-access-4clj7","readOnly":true,"mountPath":"/var/run/secrets/kubernetes.io/serviceaccount"}],"livenessProbe":{"httpGet":{"path":"/healthz","port":1936,"host":"localhost","scheme":"HTTP"},"timeoutSeconds":1,"periodSeconds":10,"successThreshold":1,"failureThreshold":3,"terminationGracePeriodSeconds":10},"readinessProbe":{"httpGet":{"path":"/healthz/ready","port":1936,"host":"localhost","scheme":"HTTP"},"timeoutSeconds":1,"periodSeconds":10,"successThreshold":1,"failureThreshold":3},"startupProbe":{"httpGet":{"path":"/healthz/ready","port":1936,"host":"localhost","scheme":"HTTP"},"timeoutSeconds":1,"periodSeconds":1,"successThreshold":1,"failureThreshold":120},"terminationMessagePath":"/dev/termination-log","terminationMessagePolicy":"FallbackToLogsOnError","imagePullPolicy":"IfNotPresent","securityContext":{"capabilities":{"drop":["KILL","MKNOD","SETGID","SETUID"]},"runAsUser":1000660000,"runAsNonRoot":true,"readOnlyRootFilesystem":false,"allowPrivilegeEscalation":true}}],"restartPolicy":"Always","terminationGracePeriodSeconds":3600,"dnsPolicy":"ClusterFirstWithHostNet","nodeSelector":{"kubernetes.io/os":"linux","node-role.kubernetes.io/master":""},"serviceAccountName":"router","serviceAccount":"router","hostNetwork":true,"securityContext":{"seLinuxOptions":{"level":"s0:c26,c5"},"supplementalGroups":[1000660000],"fsGroup":1000660000},"schedulerName":"default-scheduler","tolerations":[{"key":"kubernetes.io/e2e-evict-taint-key","operator":"Equal","value":"evictTaintVal","effect":"NoExecute"},{"key":"node.kubernetes.io/not-ready","operator":"Exists","effect":"NoExecute","tolerationSeconds":300},{"key":"node.kubernetes.io/unreachable","operator":"Exists","effect":"NoExecute","tolerationSeconds":300},{"key":"node.kubernetes.io/memory-pressure","operator":"Exists","effect":"NoSchedule"}],"priorityClassName":"system-cluster-critical","priority":2000000000,"enableServiceLinks":true,"preemptionPolicy":"PreemptLowerPriority","topologySpreadConstraints":[{"maxSkew":1,"topologyKey":"topology.kubernetes.io/zone","whenUnsatisfiable":"ScheduleAnyway","labelSelector":{"matchExpressions":[{"key":"ingresscontroller.operator.openshift.io/hash","operator":"In","values":["85fc95cf79"]}]}}]},"status":{"phase":"Pending","conditions":[{"type":"PodScheduled","status":"False","lastProbeTime":null,"lastTransitionTime":"2025-12-04T11:38:21Z","reason":"Unschedulable","message":"0/1 nodes are available: 1 node(s) had untolerated taint {node-role.kubernetes.io/master: }. preemption: 0/1 nodes are available: 1 Preemption is not helpful for scheduling."}],"qosClass":"Burstable"}}kube-rbac-proxy_current.log0000640000000000000000000000156615114274105032360 0ustar0000000000000000config/pod/openshift-ingress-operator/logs/ingress-operator-8649c48786-cx2b22025-12-04T11:38:19.925935817Z W1204 11:38:19.925582 1 deprecated.go:66] 2025-12-04T11:38:19.925935817Z ==== Removed Flag Warning ====================== 2025-12-04T11:38:19.925935817Z 2025-12-04T11:38:19.925935817Z logtostderr is removed in the k8s upstream and has no effect any more. 2025-12-04T11:38:19.925935817Z 2025-12-04T11:38:19.925935817Z =============================================== 2025-12-04T11:38:19.925935817Z 2025-12-04T11:38:19.926238927Z I1204 11:38:19.926202 1 kube-rbac-proxy.go:233] Valid token audiences: 2025-12-04T11:38:19.927754116Z I1204 11:38:19.927711 1 kube-rbac-proxy.go:347] Reading certificate files 2025-12-04T11:38:19.928158359Z I1204 11:38:19.928097 1 kube-rbac-proxy.go:395] Starting TCP socket on :9393 2025-12-04T11:38:19.928542741Z I1204 11:38:19.928514 1 kube-rbac-proxy.go:402] Listening securely on :9393 ingress-operator_current.log0000640000000000000000000005236115114274105032650 0ustar0000000000000000config/pod/openshift-ingress-operator/logs/ingress-operator-8649c48786-cx2b22025-12-04T11:59:30.889735318Z 2025-12-04T11:59:30.889Z INFO operator.init controller/controller.go:231 Starting EventSource {"controller": "dns_controller", "source": "kind source: *v1.Secret"} 2025-12-04T11:59:30.889743478Z 2025-12-04T11:59:30.889Z INFO operator.init controller/controller.go:231 Starting EventSource {"controller": "monitoring_dashboard_controller", "source": "kind source: *v1.Infrastructure"} 2025-12-04T11:59:30.889743478Z 2025-12-04T11:59:30.889Z INFO operator.init controller/controller.go:231 Starting Controller {"controller": "dns_controller"} 2025-12-04T11:59:30.889751309Z 2025-12-04T11:59:30.889Z INFO operator.init controller/controller.go:231 Starting Controller {"controller": "monitoring_dashboard_controller"} 2025-12-04T11:59:30.889758529Z 2025-12-04T11:59:30.889Z INFO operator.init controller/controller.go:231 Starting EventSource {"controller": "gatewayapi_upgradeable_controller", "source": "kind source: *v1.CustomResourceDefinition"} 2025-12-04T11:59:30.889783430Z 2025-12-04T11:59:30.889Z INFO operator.init controller/controller.go:231 Starting EventSource {"controller": "gatewayapi_upgradeable_controller", "source": "kind source: *v1.ConfigMap"} 2025-12-04T11:59:30.889790860Z 2025-12-04T11:59:30.889Z INFO operator.init controller/controller.go:231 Starting Controller {"controller": "gatewayapi_upgradeable_controller"} 2025-12-04T11:59:30.889853662Z 2025-12-04T11:59:30.889Z INFO operator.init controller/controller.go:231 Starting EventSource {"controller": "clientca_configmap_controller", "source": "kind source: *v1.IngressController"} 2025-12-04T11:59:30.889853662Z 2025-12-04T11:59:30.889Z INFO operator.init controller/controller.go:231 Starting EventSource {"controller": "canary_controller", "source": "kind source: *v1.IngressController"} 2025-12-04T11:59:30.889865162Z 2025-12-04T11:59:30.889Z INFO operator.init controller/controller.go:231 Starting EventSource {"controller": "clientca_configmap_controller", "source": "kind source: *v1.ConfigMap"} 2025-12-04T11:59:30.889865162Z 2025-12-04T11:59:30.889Z INFO operator.init controller/controller.go:231 Starting EventSource {"controller": "clientca_configmap_controller", "source": "kind source: *v1.ConfigMap"} 2025-12-04T11:59:30.889877803Z 2025-12-04T11:59:30.889Z INFO operator.init controller/controller.go:231 Starting Controller {"controller": "clientca_configmap_controller"} 2025-12-04T11:59:30.889877803Z 2025-12-04T11:59:30.889Z INFO operator.init controller/controller.go:231 Starting EventSource {"controller": "canary_controller", "source": "kind source: *v1.Route"} 2025-12-04T11:59:30.889900263Z 2025-12-04T11:59:30.889Z INFO operator.init controller/controller.go:231 Starting EventSource {"controller": "canary_controller", "source": "kind source: *v1.DaemonSet"} 2025-12-04T11:59:30.889907784Z 2025-12-04T11:59:30.889Z INFO operator.init controller/controller.go:231 Starting EventSource {"controller": "crl", "source": "informer source: 0xc0007bef98"} 2025-12-04T11:59:30.889941455Z 2025-12-04T11:59:30.889Z INFO operator.init controller/controller.go:231 Starting EventSource {"controller": "canary_controller", "source": "kind source: *v1.Service"} 2025-12-04T11:59:30.889941455Z 2025-12-04T11:59:30.889Z INFO operator.init controller/controller.go:231 Starting EventSource {"controller": "gatewayapi_controller", "source": "kind source: *v1.FeatureGate"} 2025-12-04T11:59:30.889968326Z 2025-12-04T11:59:30.889Z INFO operator.init controller/controller.go:231 Starting EventSource {"controller": "route_metrics_controller", "source": "kind source: *v1.IngressController"} 2025-12-04T11:59:30.889968326Z 2025-12-04T11:59:30.889Z INFO operator.init controller/controller.go:231 Starting EventSource {"controller": "gatewayapi_controller", "source": "kind source: *v1.CustomResourceDefinition"} 2025-12-04T11:59:30.889986606Z 2025-12-04T11:59:30.889Z INFO operator.init controller/controller.go:231 Starting EventSource {"controller": "gatewayapi_controller", "source": "kind source: *v1.CustomResourceDefinition"} 2025-12-04T11:59:30.890005337Z 2025-12-04T11:59:30.889Z INFO operator.init controller/controller.go:231 Starting EventSource {"controller": "gatewayapi_controller", "source": "kind source: *v1.CustomResourceDefinition"} 2025-12-04T11:59:30.890005337Z 2025-12-04T11:59:30.889Z INFO operator.init controller/controller.go:231 Starting Controller {"controller": "canary_controller"} 2025-12-04T11:59:30.890028557Z 2025-12-04T11:59:30.889Z INFO operator.init controller/controller.go:231 Starting EventSource {"controller": "route_metrics_controller", "source": "kind source: *v1.Route"} 2025-12-04T11:59:30.890037398Z 2025-12-04T11:59:30.890Z INFO operator.init controller/controller.go:231 Starting EventSource {"controller": "gatewayapi_controller", "source": "kind source: *v1.CustomResourceDefinition"} 2025-12-04T11:59:30.890046468Z 2025-12-04T11:59:30.890Z INFO operator.init controller/controller.go:231 Starting Controller {"controller": "route_metrics_controller"} 2025-12-04T11:59:30.890053948Z 2025-12-04T11:59:30.890Z INFO operator.init controller/controller.go:231 Starting Controller {"controller": "gatewayapi_controller"} 2025-12-04T11:59:30.890081889Z 2025-12-04T11:59:30.890Z INFO operator.init controller/controller.go:231 Starting EventSource {"controller": "crl", "source": "informer source: 0xc0007bef98"} 2025-12-04T11:59:30.890284516Z 2025-12-04T11:59:30.890Z INFO operator.init controller/controller.go:231 Starting EventSource {"controller": "crl", "source": "kind source: *v1.IngressController"} 2025-12-04T11:59:30.890284516Z 2025-12-04T11:59:30.890Z INFO operator.init controller/controller.go:231 Starting Controller {"controller": "crl"} 2025-12-04T11:59:30.905637202Z 2025-12-04T11:59:30.905Z INFO operator.certificate_publisher_controller handler/enqueue_mapped.go:119 queueing ingresscontroller {"name": "default"} 2025-12-04T11:59:30.992759202Z 2025-12-04T11:59:30.992Z INFO operator.ingress_controller handler/enqueue_mapped.go:119 queueing ingresscontroller {"name": "default", "related": ""} 2025-12-04T11:59:30.992857875Z 2025-12-04T11:59:30.992Z INFO operator.ingress_controller handler/enqueue_mapped.go:119 queueing ingresscontroller {"name": "default", "related": ""} 2025-12-04T11:59:30.998695580Z 2025-12-04T11:59:30.998Z INFO operator.ingress_controller handler/enqueue_mapped.go:119 queueing ingresscontroller {"name": "default", "related": ""} 2025-12-04T11:59:31.006612410Z 2025-12-04T11:59:31.006Z INFO operator.ingress_controller handler/enqueue_mapped.go:119 queueing ingresscontroller {"name": "default", "related": ""} 2025-12-04T11:59:31.206647905Z 2025-12-04T11:59:31.206Z INFO operator.init controller/controller.go:231 Starting workers {"controller": "certificate_controller", "worker count": 1} 2025-12-04T11:59:31.206994896Z 2025-12-04T11:59:31.206Z INFO operator.certificate_controller controller/controller.go:116 Reconciling {"request": {"name":"default","namespace":"openshift-ingress-operator"}} 2025-12-04T11:59:31.299387213Z 2025-12-04T11:59:31.299Z INFO operator.init controller/controller.go:231 Starting workers {"controller": "error_page_configmap_controller", "worker count": 1} 2025-12-04T11:59:31.395541778Z 2025-12-04T11:59:31.395Z INFO operator.ingress_controller handler/enqueue_mapped.go:119 queueing ingress {"name": "default", "related": ""} 2025-12-04T11:59:31.396820188Z 2025-12-04T11:59:31.396Z INFO operator.ingress_controller handler/enqueue_mapped.go:119 queueing ingress {"name": "default", "related": ""} 2025-12-04T11:59:31.592668231Z 2025-12-04T11:59:31.592Z INFO operator.init controller/controller.go:231 Starting workers {"controller": "route_metrics_controller", "worker count": 1} 2025-12-04T11:59:31.592749213Z 2025-12-04T11:59:31.592Z INFO operator.route_metrics_controller controller/controller.go:116 reconciling {"request": {"name":"default","namespace":"openshift-ingress-operator"}} 2025-12-04T11:59:31.593054313Z 2025-12-04T11:59:31.592Z INFO operator.init controller/controller.go:231 Starting workers {"controller": "status_controller", "worker count": 1} 2025-12-04T11:59:31.593274290Z 2025-12-04T11:59:31.593Z INFO operator.status_controller controller/controller.go:116 Reconciling {"request": {"name":"default","namespace":"openshift-ingress-operator"}} 2025-12-04T11:59:31.594112147Z 2025-12-04T11:59:31.593Z INFO operator.init controller/controller.go:231 Starting workers {"controller": "gatewayapi_upgradeable_controller", "worker count": 1} 2025-12-04T11:59:31.594112147Z 2025-12-04T11:59:31.593Z INFO operator.init controller/controller.go:231 Starting workers {"controller": "crl", "worker count": 1} 2025-12-04T11:59:31.594112147Z 2025-12-04T11:59:31.593Z INFO operator.init controller/controller.go:231 Starting workers {"controller": "clientca_configmap_controller", "worker count": 1} 2025-12-04T11:59:31.594112147Z 2025-12-04T11:59:31.594Z INFO operator.gatewayapi_upgradeable_controller controller/controller.go:116 reconciling {"request": {"name":"admin-gates","namespace":"openshift-config-managed"}} 2025-12-04T11:59:31.595583893Z 2025-12-04T11:59:31.595Z INFO operator.init controller/controller.go:231 Starting workers {"controller": "certificate_publisher_controller", "worker count": 1} 2025-12-04T11:59:31.595583893Z 2025-12-04T11:59:31.595Z INFO operator.init controller/controller.go:231 Starting workers {"controller": "gatewayapi_controller", "worker count": 1} 2025-12-04T11:59:31.595668326Z 2025-12-04T11:59:31.595Z INFO operator.certificate_publisher_controller controller/controller.go:116 Reconciling {"request": {"name":"default","namespace":"openshift-ingress-operator"}} 2025-12-04T11:59:31.595668326Z 2025-12-04T11:59:31.595Z INFO operator.gatewayapi_controller controller/controller.go:116 reconciling {"request": {"name":"cluster"}} 2025-12-04T11:59:31.598010130Z 2025-12-04T11:59:31.596Z INFO operator.init controller/controller.go:231 Starting workers {"controller": "monitoring_dashboard_controller", "worker count": 1} 2025-12-04T11:59:31.608200783Z 2025-12-04T11:59:31.608Z INFO operator.init controller/controller.go:231 Starting workers {"controller": "ingressclass_controller", "worker count": 1} 2025-12-04T11:59:31.608242605Z 2025-12-04T11:59:31.608Z INFO operator.ingressclass_controller controller/controller.go:116 reconciling {"request": {"name":"default","namespace":"openshift-ingress-operator"}} 2025-12-04T11:59:31.618985705Z 2025-12-04T11:59:31.618Z INFO operator.init controller/controller.go:231 Starting workers {"controller": "configurable_route_controller", "worker count": 1} 2025-12-04T11:59:31.619083108Z 2025-12-04T11:59:31.619Z INFO operator.configurable_route_controller controller/controller.go:116 reconciling {"request": {"name":"cluster"}} 2025-12-04T11:59:31.619206022Z 2025-12-04T11:59:31.619Z INFO operator.init controller/controller.go:231 Starting workers {"controller": "canary_controller", "worker count": 1} 2025-12-04T11:59:31.664893899Z 2025-12-04T11:59:31.664Z INFO operator.canary_controller canary/service.go:29 updated canary service {"namespace": "openshift-ingress-canary", "name": "ingress-canary", "diff": " &v1.Service{\n \tTypeMeta: {},\n \tObjectMeta: v1.ObjectMeta{\n \t\t... // 9 identical fields\n \t\tDeletionGracePeriodSeconds: nil,\n \t\tLabels: {\"ingress.openshift.io/canary\": \"canary_controller\"},\n \t\tAnnotations: map[string]string{\n- \t\t\t\"service.alpha.openshift.io/serving-cert-signed-by\": \"openshift-service-serving-signer@1764848266\",\n \t\t\t\"service.beta.openshift.io/serving-cert-secret-name\": \"canary-serving-cert\",\n- \t\t\t\"service.beta.openshift.io/serving-cert-signed-by\": \"openshift-service-serving-signer@1764848266\",\n \t\t},\n \t\tOwnerReferences: {{APIVersion: \"apps/v1\", Kind: \"daemonset\", Name: \"ingress-canary\", UID: \"bddf607e-df5e-4d58-8f8a-44488bd376c8\", ...}},\n \t\tFinalizers: nil,\n \t\tManagedFields: {{Manager: \"ingress-operator\", Operation: \"Update\", APIVersion: \"v1\", Time: s\"2025-12-04 11:47:24 +0000 UTC\", ...}, {Manager: \"service-ca-operator\", Operation: \"Update\", APIVersion: \"v1\", Time: s\"2025-12-04 11:50:06 +0000 UTC\", ...}},\n \t},\n \tSpec: {Ports: {{Name: \"8443-tcp\", Protocol: \"TCP\", Port: 8443, TargetPort: {IntVal: 8443}, ...}, {Name: \"8888-tcp\", Protocol: \"TCP\", Port: 8888, TargetPort: {IntVal: 8888}, ...}}, Selector: {\"ingresscanary.operator.openshift.io/daemonset-ingresscanary\": \"canary_controller\"}, ClusterIP: \"172.30.249.0\", ClusterIPs: {\"172.30.249.0\"}, ...},\n \tStatus: {},\n }\n"} 2025-12-04T11:59:31.705830115Z 2025-12-04T11:59:31.705Z INFO operator.canary_controller canary/route.go:35 created canary route {"namespace": "openshift-ingress-canary", "name": "canary"} 2025-12-04T11:59:31.714535931Z 2025-12-04T11:59:31.714Z INFO operator.init controller/controller.go:231 Starting workers {"controller": "dns_controller", "worker count": 1} 2025-12-04T11:59:31.714820530Z 2025-12-04T11:59:31.714Z INFO operator.init controller/controller.go:231 Starting workers {"controller": "ingress_controller", "worker count": 1} 2025-12-04T11:59:31.715319246Z 2025-12-04T11:59:31.715Z INFO operator.ingress_controller controller/controller.go:116 reconciling {"request": {"name":"default","namespace":"openshift-ingress-operator"}} 2025-12-04T11:59:31.747091312Z 2025-12-04T11:59:31.746Z INFO operator.certificate_controller controller/controller.go:116 Reconciling {"request": {"name":"default","namespace":"openshift-ingress-operator"}} 2025-12-04T11:59:31.747091312Z 2025-12-04T11:59:31.747Z INFO operator.status_controller controller/controller.go:116 Reconciling {"request": {"name":"default","namespace":"openshift-ingress-operator"}} 2025-12-04T11:59:31.747438593Z 2025-12-04T11:59:31.747Z INFO operator.ingressclass_controller controller/controller.go:116 reconciling {"request": {"name":"default","namespace":"openshift-ingress-operator"}} 2025-12-04T11:59:31.809234790Z 2025-12-04T11:59:31.809Z INFO operator.route_metrics_controller controller/controller.go:116 reconciling {"request": {"name":"default","namespace":"openshift-ingress-operator"}} 2025-12-04T11:59:31.940984363Z 2025-12-04T11:59:31.940Z ERROR operator.ingress_controller controller/controller.go:116 got retryable error; requeueing {"after": "38m49.061835426s", "error": "IngressController may become degraded soon: DeploymentReplicasAllAvailable=False"} 2025-12-04T11:59:31.940984363Z 2025-12-04T11:59:31.940Z INFO operator.ingress_controller controller/controller.go:116 reconciling {"request": {"name":"default","namespace":"openshift-ingress-operator"}} 2025-12-04T11:59:32.019792909Z 2025-12-04T11:59:32.019Z ERROR operator.ingress_controller controller/controller.go:116 got retryable error; requeueing {"after": "58.981830463s", "error": "IngressController may become degraded soon: DeploymentReplicasAllAvailable=False, CanaryChecksSucceeding=Unknown"} 2025-12-04T11:59:49.752115482Z 2025-12-04T11:59:49.751Z INFO operator.ingress_controller handler/enqueue_mapped.go:119 queueing ingresscontroller {"name": "default", "related": ""} 2025-12-04T11:59:49.752115482Z 2025-12-04T11:59:49.752Z INFO operator.ingress_controller handler/enqueue_mapped.go:119 queueing ingresscontroller {"name": "default", "related": ""} 2025-12-04T11:59:49.752115482Z 2025-12-04T11:59:49.752Z INFO operator.ingress_controller controller/controller.go:116 reconciling {"request": {"name":"default","namespace":"openshift-ingress-operator"}} 2025-12-04T11:59:49.752396191Z 2025-12-04T11:59:49.751Z INFO operator.configurable_route_controller controller/controller.go:116 reconciling {"request": {"name":"cluster"}} 2025-12-04T11:59:49.931826964Z 2025-12-04T11:59:49.929Z ERROR operator.ingress_controller controller/controller.go:116 got retryable error; requeueing {"after": "41.071783501s", "error": "IngressController may become degraded soon: DeploymentReplicasAllAvailable=False, CanaryChecksSucceeding=Unknown"} 2025-12-04T11:59:50.081432171Z 2025-12-04T11:59:50.078Z INFO operator.ingress_controller handler/enqueue_mapped.go:119 queueing ingresscontroller {"name": "default", "related": ""} 2025-12-04T11:59:50.081432171Z 2025-12-04T11:59:50.078Z INFO operator.ingress_controller handler/enqueue_mapped.go:119 queueing ingresscontroller {"name": "default", "related": ""} 2025-12-04T11:59:50.081432171Z 2025-12-04T11:59:50.078Z INFO operator.ingress_controller controller/controller.go:116 reconciling {"request": {"name":"default","namespace":"openshift-ingress-operator"}} 2025-12-04T11:59:50.081432171Z 2025-12-04T11:59:50.078Z INFO operator.configurable_route_controller controller/controller.go:116 reconciling {"request": {"name":"cluster"}} 2025-12-04T11:59:50.176291156Z 2025-12-04T11:59:50.176Z ERROR operator.ingress_controller controller/controller.go:116 got retryable error; requeueing {"after": "40.825715938s", "error": "IngressController may become degraded soon: DeploymentReplicasAllAvailable=False, CanaryChecksSucceeding=Unknown"} 2025-12-04T11:59:53.309876779Z 2025-12-04T11:59:53.308Z INFO operator.ingress_controller handler/enqueue_mapped.go:119 queueing ingresscontroller {"name": "default", "related": ""} 2025-12-04T11:59:53.309876779Z 2025-12-04T11:59:53.309Z INFO operator.ingress_controller handler/enqueue_mapped.go:119 queueing ingresscontroller {"name": "default", "related": ""} 2025-12-04T11:59:53.309876779Z 2025-12-04T11:59:53.309Z INFO operator.ingress_controller controller/controller.go:116 reconciling {"request": {"name":"default","namespace":"openshift-ingress-operator"}} 2025-12-04T11:59:53.309876779Z 2025-12-04T11:59:53.309Z INFO operator.configurable_route_controller controller/controller.go:116 reconciling {"request": {"name":"cluster"}} 2025-12-04T11:59:53.821497062Z 2025-12-04T11:59:53.818Z INFO operator.ingress_controller handler/enqueue_mapped.go:119 queueing ingresscontroller {"name": "default", "related": ""} 2025-12-04T11:59:53.821497062Z 2025-12-04T11:59:53.818Z INFO operator.ingress_controller handler/enqueue_mapped.go:119 queueing ingresscontroller {"name": "default", "related": ""} 2025-12-04T11:59:53.821497062Z 2025-12-04T11:59:53.818Z INFO operator.configurable_route_controller controller/controller.go:116 reconciling {"request": {"name":"cluster"}} 2025-12-04T11:59:54.147624902Z 2025-12-04T11:59:54.147Z ERROR operator.ingress_controller controller/controller.go:116 got retryable error; requeueing {"after": "36.857449039s", "error": "IngressController may become degraded soon: DeploymentReplicasAllAvailable=False, CanaryChecksSucceeding=Unknown"} 2025-12-04T11:59:54.147624902Z 2025-12-04T11:59:54.147Z INFO operator.ingress_controller controller/controller.go:116 reconciling {"request": {"name":"default","namespace":"openshift-ingress-operator"}} 2025-12-04T11:59:54.492358600Z 2025-12-04T11:59:54.492Z ERROR operator.ingress_controller controller/controller.go:116 got retryable error; requeueing {"after": "36.509711247s", "error": "IngressController may become degraded soon: DeploymentReplicasAllAvailable=False, CanaryChecksSucceeding=Unknown"} 2025-12-04T12:00:31.002082710Z 2025-12-04T12:00:31.001Z INFO operator.ingress_controller controller/controller.go:116 reconciling {"request": {"name":"default","namespace":"openshift-ingress-operator"}} 2025-12-04T12:00:31.136563879Z 2025-12-04T12:00:31.136Z INFO operator.certificate_controller controller/controller.go:116 Reconciling {"request": {"name":"default","namespace":"openshift-ingress-operator"}} 2025-12-04T12:00:31.136563879Z 2025-12-04T12:00:31.136Z INFO operator.route_metrics_controller controller/controller.go:116 reconciling {"request": {"name":"default","namespace":"openshift-ingress-operator"}} 2025-12-04T12:00:31.136597800Z 2025-12-04T12:00:31.136Z INFO operator.ingressclass_controller controller/controller.go:116 reconciling {"request": {"name":"default","namespace":"openshift-ingress-operator"}} 2025-12-04T12:00:31.137926993Z 2025-12-04T12:00:31.136Z INFO operator.status_controller controller/controller.go:116 Reconciling {"request": {"name":"default","namespace":"openshift-ingress-operator"}} 2025-12-04T12:00:31.140326179Z 2025-12-04T12:00:31.140Z ERROR operator.ingress_controller controller/controller.go:116 got retryable error; requeueing {"after": "1m0s", "error": "IngressController is degraded: CanaryChecksSucceeding=Unknown (CanaryRouteNotAdmitted: Canary route is not admitted by the default ingress controller)"} 2025-12-04T12:00:31.140399001Z 2025-12-04T12:00:31.140Z INFO operator.ingress_controller controller/controller.go:116 reconciling {"request": {"name":"default","namespace":"openshift-ingress-operator"}} 2025-12-04T12:00:31.205448952Z 2025-12-04T12:00:31.205Z INFO operator.status_controller controller/controller.go:116 Reconciling {"request": {"name":"default","namespace":"openshift-ingress-operator"}} 2025-12-04T12:00:31.569071928Z 2025-12-04T12:00:31.568Z ERROR operator.ingress_controller controller/controller.go:116 got retryable error; requeueing {"after": "1m0s", "error": "IngressController is degraded: CanaryChecksSucceeding=Unknown (CanaryRouteNotAdmitted: Canary route is not admitted by the default ingress controller)"} 2025-12-04T12:01:31.141485668Z 2025-12-04T12:01:31.141Z INFO operator.ingress_controller controller/controller.go:116 reconciling {"request": {"name":"default","namespace":"openshift-ingress-operator"}} 2025-12-04T12:01:31.226033514Z 2025-12-04T12:01:31.225Z ERROR operator.ingress_controller controller/controller.go:116 got retryable error; requeueing {"after": "1m0s", "error": "IngressController is degraded: CanaryChecksSucceeding=Unknown (CanaryRouteNotAdmitted: Canary route is not admitted by the default ingress controller)"} ingress-operator_previous.log0000640000000000000000000006534115114274105033044 0ustar0000000000000000config/pod/openshift-ingress-operator/logs/ingress-operator-8649c48786-cx2b22025-12-04T11:56:19.278052168Z 2025-12-04T11:56:19.277Z ERROR operator.init.controller-runtime.source.EventHandler wait/loop.go:87 failed to get informer from cache {"error": "failed to get API group resources: unable to retrieve the complete list of server APIs: route.openshift.io/v1: the server is currently unable to handle the request"} 2025-12-04T11:56:19.278052168Z 2025-12-04T11:56:19.277Z ERROR operator.init.controller-runtime.source.EventHandler wait/loop.go:87 failed to get informer from cache {"error": "failed to get restmapping: failed to get API group resources: unable to retrieve the complete list of server APIs: route.openshift.io/v1: the server is currently unable to handle the request"} 2025-12-04T11:56:26.477120531Z 2025-12-04T11:56:26.476Z ERROR operator.ingress_controller controller/controller.go:116 got retryable error; requeueing {"after": "41m54.525271362s", "error": "IngressController may become degraded soon: DeploymentReplicasAllAvailable=False"} 2025-12-04T11:56:29.277088130Z 2025-12-04T11:56:29.276Z ERROR operator.init.controller-runtime.source.EventHandler wait/loop.go:87 failed to get informer from cache {"error": "failed to get restmapping: failed to get API group resources: unable to retrieve the complete list of server APIs: route.openshift.io/v1: the server is currently unable to handle the request"} 2025-12-04T11:56:29.277145542Z 2025-12-04T11:56:29.277Z ERROR operator.init.controller-runtime.source.EventHandler wait/loop.go:87 failed to get informer from cache {"error": "failed to get API group resources: unable to retrieve the complete list of server APIs: route.openshift.io/v1: the server is currently unable to handle the request"} 2025-12-04T11:56:39.277373488Z 2025-12-04T11:56:39.277Z ERROR operator.init.controller-runtime.source.EventHandler wait/loop.go:87 failed to get informer from cache {"error": "failed to get API group resources: unable to retrieve the complete list of server APIs: route.openshift.io/v1: the server is currently unable to handle the request"} 2025-12-04T11:56:39.277506793Z 2025-12-04T11:56:39.277Z ERROR operator.init.controller-runtime.source.EventHandler wait/loop.go:87 failed to get informer from cache {"error": "failed to get restmapping: failed to get API group resources: unable to retrieve the complete list of server APIs: route.openshift.io/v1: the server is currently unable to handle the request"} 2025-12-04T11:56:49.277287707Z 2025-12-04T11:56:49.277Z ERROR operator.init.controller-runtime.source.EventHandler wait/loop.go:87 failed to get informer from cache {"error": "failed to get API group resources: unable to retrieve the complete list of server APIs: route.openshift.io/v1: the server is currently unable to handle the request"} 2025-12-04T11:56:49.278067181Z 2025-12-04T11:56:49.278Z ERROR operator.init.controller-runtime.source.EventHandler wait/loop.go:87 failed to get informer from cache {"error": "failed to get restmapping: failed to get API group resources: unable to retrieve the complete list of server APIs: route.openshift.io/v1: the server is currently unable to handle the request"} 2025-12-04T11:56:59.277454508Z 2025-12-04T11:56:59.277Z ERROR operator.init.controller-runtime.source.EventHandler wait/loop.go:87 failed to get informer from cache {"error": "failed to get restmapping: failed to get API group resources: unable to retrieve the complete list of server APIs: route.openshift.io/v1: the server is currently unable to handle the request"} 2025-12-04T11:56:59.277454508Z 2025-12-04T11:56:59.277Z ERROR operator.init.controller-runtime.source.EventHandler wait/loop.go:87 failed to get informer from cache {"error": "failed to get API group resources: unable to retrieve the complete list of server APIs: route.openshift.io/v1: the server is currently unable to handle the request"} 2025-12-04T11:57:09.277527288Z 2025-12-04T11:57:09.277Z ERROR operator.init.controller-runtime.source.EventHandler wait/loop.go:87 failed to get informer from cache {"error": "failed to get restmapping: failed to get API group resources: unable to retrieve the complete list of server APIs: route.openshift.io/v1: the server is currently unable to handle the request"} 2025-12-04T11:57:09.277702473Z 2025-12-04T11:57:09.277Z ERROR operator.init.controller-runtime.source.EventHandler wait/loop.go:87 failed to get informer from cache {"error": "failed to get API group resources: unable to retrieve the complete list of server APIs: route.openshift.io/v1: the server is currently unable to handle the request"} 2025-12-04T11:57:19.278520492Z 2025-12-04T11:57:19.278Z ERROR operator.init.controller-runtime.source.EventHandler wait/loop.go:87 failed to get informer from cache {"error": "failed to get API group resources: unable to retrieve the complete list of server APIs: route.openshift.io/v1: the server is currently unable to handle the request"} 2025-12-04T11:57:19.278575794Z 2025-12-04T11:57:19.278Z ERROR operator.init.controller-runtime.source.EventHandler wait/loop.go:87 failed to get informer from cache {"error": "failed to get restmapping: failed to get API group resources: unable to retrieve the complete list of server APIs: route.openshift.io/v1: the server is currently unable to handle the request"} 2025-12-04T11:57:29.277197991Z 2025-12-04T11:57:29.277Z ERROR operator.init.controller-runtime.source.EventHandler wait/loop.go:87 failed to get informer from cache {"error": "failed to get restmapping: failed to get API group resources: unable to retrieve the complete list of server APIs: route.openshift.io/v1: the server is currently unable to handle the request"} 2025-12-04T11:57:29.277329015Z 2025-12-04T11:57:29.277Z ERROR operator.init.controller-runtime.source.EventHandler wait/loop.go:87 failed to get informer from cache {"error": "failed to get API group resources: unable to retrieve the complete list of server APIs: route.openshift.io/v1: the server is currently unable to handle the request"} 2025-12-04T11:57:39.278499442Z 2025-12-04T11:57:39.278Z ERROR operator.init.controller-runtime.source.EventHandler wait/loop.go:87 failed to get informer from cache {"error": "failed to get restmapping: failed to get API group resources: unable to retrieve the complete list of server APIs: route.openshift.io/v1: the server is currently unable to handle the request"} 2025-12-04T11:57:39.278499442Z 2025-12-04T11:57:39.278Z ERROR operator.init.controller-runtime.source.EventHandler wait/loop.go:87 failed to get informer from cache {"error": "failed to get API group resources: unable to retrieve the complete list of server APIs: route.openshift.io/v1: the server is currently unable to handle the request"} 2025-12-04T11:57:39.680591248Z 2025-12-04T11:57:39.680Z ERROR operator.init controller/controller.go:205 Could not wait for Cache to sync {"controller": "canary_controller", "error": "failed to wait for canary_controller caches to sync: timed out waiting for cache to be synced for Kind *v1.Route"} 2025-12-04T11:57:39.680591248Z 2025-12-04T11:57:39.680Z INFO operator.init runtime/asm_amd64.s:1695 Stopping and waiting for non leader election runnables 2025-12-04T11:57:39.680591248Z 2025-12-04T11:57:39.680Z INFO operator.init runtime/asm_amd64.s:1695 Stopping and waiting for leader election runnables 2025-12-04T11:57:39.680653309Z 2025-12-04T11:57:39.680Z INFO operator.init manager/runnable_group.go:226 Shutdown signal received, waiting for all workers to finish {"controller": "configurable_route_controller"} 2025-12-04T11:57:39.680653309Z 2025-12-04T11:57:39.680Z INFO operator.init manager/runnable_group.go:226 Shutdown signal received, waiting for all workers to finish {"controller": "ingress_controller"} 2025-12-04T11:57:39.680682380Z 2025-12-04T11:57:39.680Z INFO operator.init manager/runnable_group.go:226 Shutdown signal received, waiting for all workers to finish {"controller": "clientca_configmap_controller"} 2025-12-04T11:57:39.680682380Z 2025-12-04T11:57:39.680Z INFO operator.init manager/runnable_group.go:226 Shutdown signal received, waiting for all workers to finish {"controller": "certificate_publisher_controller"} 2025-12-04T11:57:39.680711491Z 2025-12-04T11:57:39.680Z INFO operator.init manager/runnable_group.go:226 Shutdown signal received, waiting for all workers to finish {"controller": "dns_controller"} 2025-12-04T11:57:39.680711491Z 2025-12-04T11:57:39.680Z INFO operator.init manager/runnable_group.go:226 Shutdown signal received, waiting for all workers to finish {"controller": "gatewayapi_controller"} 2025-12-04T11:57:39.680725401Z 2025-12-04T11:57:39.680Z INFO operator.init manager/runnable_group.go:226 Shutdown signal received, waiting for all workers to finish {"controller": "gatewayapi_upgradeable_controller"} 2025-12-04T11:57:39.680798313Z 2025-12-04T11:57:39.680Z INFO operator.init controller/controller.go:231 Starting workers {"controller": "route_metrics_controller", "worker count": 1} 2025-12-04T11:57:39.680798313Z 2025-12-04T11:57:39.680Z INFO operator.init manager/runnable_group.go:226 Shutdown signal received, waiting for all workers to finish {"controller": "route_metrics_controller"} 2025-12-04T11:57:39.680854465Z 2025-12-04T11:57:39.680Z INFO operator.init manager/runnable_group.go:226 Shutdown signal received, waiting for all workers to finish {"controller": "ingressclass_controller"} 2025-12-04T11:57:39.680863855Z 2025-12-04T11:57:39.680Z INFO operator.init manager/runnable_group.go:226 Shutdown signal received, waiting for all workers to finish {"controller": "crl"} 2025-12-04T11:57:39.680873065Z 2025-12-04T11:57:39.680Z INFO operator.init manager/runnable_group.go:226 Shutdown signal received, waiting for all workers to finish {"controller": "error_page_configmap_controller"} 2025-12-04T11:57:39.680884206Z 2025-12-04T11:57:39.680Z INFO operator.init manager/runnable_group.go:226 Shutdown signal received, waiting for all workers to finish {"controller": "status_controller"} 2025-12-04T11:57:39.680908496Z 2025-12-04T11:57:39.680Z INFO operator.init manager/runnable_group.go:226 Shutdown signal received, waiting for all workers to finish {"controller": "monitoring_dashboard_controller"} 2025-12-04T11:57:39.680918557Z 2025-12-04T11:57:39.680Z INFO operator.init manager/runnable_group.go:226 Shutdown signal received, waiting for all workers to finish {"controller": "certificate_controller"} 2025-12-04T11:57:39.680957888Z 2025-12-04T11:57:39.680Z INFO operator.init manager/runnable_group.go:226 All workers finished {"controller": "certificate_publisher_controller"} 2025-12-04T11:57:39.680967818Z 2025-12-04T11:57:39.680Z INFO operator.route_metrics_controller controller/controller.go:116 reconciling {"request": {"name":"default","namespace":"openshift-ingress-operator"}} 2025-12-04T11:57:39.680976188Z 2025-12-04T11:57:39.680Z INFO operator.init manager/runnable_group.go:226 All workers finished {"controller": "gatewayapi_controller"} 2025-12-04T11:57:39.680987718Z 2025-12-04T11:57:39.680Z INFO operator.init manager/runnable_group.go:226 All workers finished {"controller": "configurable_route_controller"} 2025-12-04T11:57:39.680998379Z 2025-12-04T11:57:39.680Z INFO operator.init manager/runnable_group.go:226 All workers finished {"controller": "gatewayapi_upgradeable_controller"} 2025-12-04T11:57:39.681014609Z 2025-12-04T11:57:39.680Z INFO operator.init manager/runnable_group.go:226 All workers finished {"controller": "monitoring_dashboard_controller"} 2025-12-04T11:57:39.681014609Z 2025-12-04T11:57:39.680Z INFO operator.init manager/runnable_group.go:226 All workers finished {"controller": "status_controller"} 2025-12-04T11:57:39.681022859Z 2025-12-04T11:57:39.681Z INFO operator.init manager/runnable_group.go:226 All workers finished {"controller": "ingressclass_controller"} 2025-12-04T11:57:39.681032470Z 2025-12-04T11:57:39.681Z INFO operator.init manager/runnable_group.go:226 All workers finished {"controller": "error_page_configmap_controller"} 2025-12-04T11:57:39.681040980Z 2025-12-04T11:57:39.681Z INFO operator.init manager/runnable_group.go:226 All workers finished {"controller": "dns_controller"} 2025-12-04T11:57:39.681069191Z 2025-12-04T11:57:39.681Z INFO operator.init manager/runnable_group.go:226 All workers finished {"controller": "ingress_controller"} 2025-12-04T11:57:39.681069191Z 2025-12-04T11:57:39.681Z INFO operator.init manager/runnable_group.go:226 All workers finished {"controller": "clientca_configmap_controller"} 2025-12-04T11:57:39.681069191Z 2025-12-04T11:57:39.681Z INFO operator.init manager/runnable_group.go:226 All workers finished {"controller": "crl"} 2025-12-04T11:57:39.681116422Z 2025-12-04T11:57:39.681Z INFO operator.init manager/runnable_group.go:226 All workers finished {"controller": "certificate_controller"} 2025-12-04T11:57:39.683504655Z 2025-12-04T11:57:39.683Z ERROR operator.init controller/controller.go:263 Reconciler error {"controller": "route_metrics_controller", "object": {"name":"default","namespace":"openshift-ingress-operator"}, "namespace": "openshift-ingress-operator", "name": "default", "reconcileID": "6d676a72-3b90-410a-b017-ad8f2a8631f5", "error": "failed to get Ingress Controller \"openshift-ingress-operator/default\": Timeout: failed waiting for *v1.IngressController Informer to sync"} 2025-12-04T11:57:39.683527046Z 2025-12-04T11:57:39.683Z INFO operator.init manager/runnable_group.go:226 All workers finished {"controller": "route_metrics_controller"} 2025-12-04T11:57:39.683562117Z 2025-12-04T11:57:39.683Z INFO operator.init runtime/asm_amd64.s:1695 Stopping and waiting for caches 2025-12-04T11:57:39.683722541Z W1204 11:57:39.683694 1 reflector.go:484] sigs.k8s.io/controller-runtime/pkg/cache/internal/informers.go:106: watch of *v1.Event ended with: an error on the server ("unable to decode an event from the watch stream: context canceled") has prevented the request from succeeding 2025-12-04T11:57:39.683750662Z W1204 11:57:39.683707 1 reflector.go:484] sigs.k8s.io/controller-runtime/pkg/cache/internal/informers.go:106: watch of *v1.Role ended with: an error on the server ("unable to decode an event from the watch stream: context canceled") has prevented the request from succeeding 2025-12-04T11:57:39.683865925Z W1204 11:57:39.683798 1 reflector.go:484] sigs.k8s.io/controller-runtime/pkg/cache/internal/informers.go:106: watch of *v1.Deployment ended with: an error on the server ("unable to decode an event from the watch stream: context canceled") has prevented the request from succeeding 2025-12-04T11:57:39.683865925Z W1204 11:57:39.683828 1 reflector.go:484] sigs.k8s.io/controller-runtime/pkg/cache/internal/informers.go:106: watch of *v1.DaemonSet ended with: an error on the server ("unable to decode an event from the watch stream: context canceled") has prevented the request from succeeding 2025-12-04T11:57:39.683921177Z W1204 11:57:39.683908 1 reflector.go:484] sigs.k8s.io/controller-runtime/pkg/cache/internal/informers.go:106: watch of *v1.Pod ended with: an error on the server ("unable to decode an event from the watch stream: context canceled") has prevented the request from succeeding 2025-12-04T11:57:39.683988499Z W1204 11:57:39.683947 1 reflector.go:484] sigs.k8s.io/controller-runtime/pkg/cache/internal/informers.go:106: watch of *v1.RoleBinding ended with: an error on the server ("unable to decode an event from the watch stream: context canceled") has prevented the request from succeeding 2025-12-04T11:57:39.684026100Z W1204 11:57:39.684010 1 reflector.go:484] sigs.k8s.io/controller-runtime/pkg/cache/internal/informers.go:106: watch of *v1.FeatureGate ended with: an error on the server ("unable to decode an event from the watch stream: context canceled") has prevented the request from succeeding 2025-12-04T11:57:39.684056410Z W1204 11:57:39.684039 1 reflector.go:484] sigs.k8s.io/controller-runtime/pkg/cache/internal/informers.go:106: watch of *v1.DaemonSet ended with: an error on the server ("unable to decode an event from the watch stream: context canceled") has prevented the request from succeeding 2025-12-04T11:57:39.684075711Z W1204 11:57:39.684060 1 reflector.go:484] sigs.k8s.io/controller-runtime/pkg/cache/internal/informers.go:106: watch of *v1.Infrastructure ended with: an error on the server ("unable to decode an event from the watch stream: context canceled") has prevented the request from succeeding 2025-12-04T11:57:39.684099182Z W1204 11:57:39.684088 1 reflector.go:484] sigs.k8s.io/controller-runtime/pkg/cache/internal/informers.go:106: watch of *v1.Service ended with: an error on the server ("unable to decode an event from the watch stream: context canceled") has prevented the request from succeeding 2025-12-04T11:57:39.684168233Z W1204 11:57:39.684142 1 reflector.go:484] sigs.k8s.io/controller-runtime/pkg/cache/internal/informers.go:106: watch of *v1.DNS ended with: an error on the server ("unable to decode an event from the watch stream: context canceled") has prevented the request from succeeding 2025-12-04T11:57:39.684168233Z W1204 11:57:39.684149 1 reflector.go:484] sigs.k8s.io/controller-runtime/pkg/cache/internal/informers.go:106: watch of *v1.Deployment ended with: an error on the server ("unable to decode an event from the watch stream: context canceled") has prevented the request from succeeding 2025-12-04T11:57:39.684200774Z W1204 11:57:39.684186 1 reflector.go:484] sigs.k8s.io/controller-runtime/pkg/cache/internal/informers.go:106: watch of *v1.Pod ended with: an error on the server ("unable to decode an event from the watch stream: context canceled") has prevented the request from succeeding 2025-12-04T11:57:39.684233315Z W1204 11:57:39.684222 1 reflector.go:484] sigs.k8s.io/controller-runtime/pkg/cache/internal/informers.go:106: watch of *v1.DaemonSet ended with: an error on the server ("unable to decode an event from the watch stream: context canceled") has prevented the request from succeeding 2025-12-04T11:57:39.684292617Z W1204 11:57:39.684270 1 reflector.go:484] sigs.k8s.io/controller-runtime/pkg/cache/internal/informers.go:106: watch of *v1.DaemonSet ended with: an error on the server ("unable to decode an event from the watch stream: context canceled") has prevented the request from succeeding 2025-12-04T11:57:39.684339798Z W1204 11:57:39.684328 1 reflector.go:484] sigs.k8s.io/controller-runtime/pkg/cache/internal/informers.go:106: watch of *v1.DNSRecord ended with: an error on the server ("unable to decode an event from the watch stream: context canceled") has prevented the request from succeeding 2025-12-04T11:57:39.684369079Z W1204 11:57:39.684349 1 reflector.go:484] sigs.k8s.io/controller-runtime/pkg/cache/internal/informers.go:106: watch of *v1.DNSRecord ended with: an error on the server ("unable to decode an event from the watch stream: context canceled") has prevented the request from succeeding 2025-12-04T11:57:39.684388679Z W1204 11:57:39.684375 1 reflector.go:484] sigs.k8s.io/controller-runtime/pkg/cache/internal/informers.go:106: watch of *v1.ConfigMap ended with: an error on the server ("unable to decode an event from the watch stream: context canceled") has prevented the request from succeeding 2025-12-04T11:57:39.684436901Z W1204 11:57:39.684420 1 reflector.go:484] sigs.k8s.io/controller-runtime/pkg/cache/internal/informers.go:106: watch of *v1.ConfigMap ended with: an error on the server ("unable to decode an event from the watch stream: context canceled") has prevented the request from succeeding 2025-12-04T11:57:39.684466822Z W1204 11:57:39.684455 1 reflector.go:484] sigs.k8s.io/controller-runtime/pkg/cache/internal/informers.go:106: watch of *v1.RoleBinding ended with: an error on the server ("unable to decode an event from the watch stream: context canceled") has prevented the request from succeeding 2025-12-04T11:57:39.684515563Z W1204 11:57:39.684501 1 reflector.go:484] sigs.k8s.io/controller-runtime/pkg/cache/internal/informers.go:106: watch of *v1.Secret ended with: an error on the server ("unable to decode an event from the watch stream: context canceled") has prevented the request from succeeding 2025-12-04T11:57:39.684539593Z W1204 11:57:39.684529 1 reflector.go:484] sigs.k8s.io/controller-runtime/pkg/cache/internal/informers.go:106: watch of *v1.Deployment ended with: an error on the server ("unable to decode an event from the watch stream: context canceled") has prevented the request from succeeding 2025-12-04T11:57:39.684567214Z W1204 11:57:39.684553 1 reflector.go:484] sigs.k8s.io/controller-runtime/pkg/cache/internal/informers.go:106: watch of *v1.Secret ended with: an error on the server ("unable to decode an event from the watch stream: context canceled") has prevented the request from succeeding 2025-12-04T11:57:39.684613956Z W1204 11:57:39.684596 1 reflector.go:484] sigs.k8s.io/controller-runtime/pkg/cache/internal/informers.go:106: watch of *v1.Secret ended with: an error on the server ("unable to decode an event from the watch stream: context canceled") has prevented the request from succeeding 2025-12-04T11:57:39.684647776Z W1204 11:57:39.684634 1 reflector.go:484] sigs.k8s.io/controller-runtime/pkg/cache/internal/informers.go:106: watch of *v1.Service ended with: an error on the server ("unable to decode an event from the watch stream: context canceled") has prevented the request from succeeding 2025-12-04T11:57:39.684704328Z W1204 11:57:39.684686 1 reflector.go:484] sigs.k8s.io/controller-runtime/pkg/cache/internal/informers.go:106: watch of *v1.IngressController ended with: an error on the server ("unable to decode an event from the watch stream: context canceled") has prevented the request from succeeding 2025-12-04T11:57:39.684756609Z W1204 11:57:39.684739 1 reflector.go:484] sigs.k8s.io/controller-runtime/pkg/cache/internal/informers.go:106: watch of *v1.Secret ended with: an error on the server ("unable to decode an event from the watch stream: context canceled") has prevented the request from succeeding 2025-12-04T11:57:39.684783270Z 2025-12-04T11:57:39.684Z INFO operator.init runtime/asm_amd64.s:1695 Stopping and waiting for webhooks 2025-12-04T11:57:39.684783270Z 2025-12-04T11:57:39.684Z INFO operator.init runtime/asm_amd64.s:1695 Stopping and waiting for HTTP servers 2025-12-04T11:57:39.684861462Z W1204 11:57:39.684847 1 reflector.go:484] sigs.k8s.io/controller-runtime/pkg/cache/internal/informers.go:106: watch of *v1.DNSRecord ended with: an error on the server ("unable to decode an event from the watch stream: context canceled") has prevented the request from succeeding 2025-12-04T11:57:39.684964115Z W1204 11:57:39.684953 1 reflector.go:484] sigs.k8s.io/controller-runtime/pkg/cache/internal/informers.go:106: watch of *v1.Pod ended with: an error on the server ("unable to decode an event from the watch stream: context canceled") has prevented the request from succeeding 2025-12-04T11:57:39.685003476Z W1204 11:57:39.684092 1 reflector.go:484] sigs.k8s.io/controller-runtime/pkg/cache/internal/informers.go:106: watch of *v1.Role ended with: an error on the server ("unable to decode an event from the watch stream: context canceled") has prevented the request from succeeding 2025-12-04T11:57:39.685089168Z W1204 11:57:39.685051 1 reflector.go:484] sigs.k8s.io/controller-runtime/pkg/cache/internal/informers.go:106: watch of *v1.Service ended with: an error on the server ("unable to decode an event from the watch stream: context canceled") has prevented the request from succeeding 2025-12-04T11:57:39.685117309Z W1204 11:57:39.684224 1 reflector.go:484] sigs.k8s.io/controller-runtime/pkg/cache/internal/informers.go:106: watch of *v1.Pod ended with: an error on the server ("unable to decode an event from the watch stream: context canceled") has prevented the request from succeeding 2025-12-04T11:57:39.685190041Z W1204 11:57:39.685170 1 reflector.go:484] sigs.k8s.io/controller-runtime/pkg/cache/internal/informers.go:106: watch of *v1.DaemonSet ended with: an error on the server ("unable to decode an event from the watch stream: context canceled") has prevented the request from succeeding 2025-12-04T11:57:39.685218352Z W1204 11:57:39.684331 1 reflector.go:484] sigs.k8s.io/controller-runtime/pkg/cache/internal/informers.go:106: watch of *v1.Service ended with: an error on the server ("unable to decode an event from the watch stream: context canceled") has prevented the request from succeeding 2025-12-04T11:57:39.685241052Z W1204 11:57:39.684467 1 reflector.go:484] sigs.k8s.io/controller-runtime/pkg/cache/internal/informers.go:106: watch of *v1.ConfigMap ended with: an error on the server ("unable to decode an event from the watch stream: context canceled") has prevented the request from succeeding 2025-12-04T11:57:39.685279364Z W1204 11:57:39.684648 1 reflector.go:484] sigs.k8s.io/controller-runtime/pkg/cache/internal/informers.go:106: watch of *v1.IngressController ended with: an error on the server ("unable to decode an event from the watch stream: context canceled") has prevented the request from succeeding 2025-12-04T11:57:39.685310133Z 2025-12-04T11:57:39.685Z INFO operator.init.controller-runtime.metrics runtime/asm_amd64.s:1695 Shutting down metrics server with timeout of 1 minute 2025-12-04T11:57:39.685318194Z W1204 11:57:39.683922 1 reflector.go:484] sigs.k8s.io/controller-runtime/pkg/cache/internal/informers.go:106: watch of *v1.Service ended with: an error on the server ("unable to decode an event from the watch stream: context canceled") has prevented the request from succeeding 2025-12-04T11:57:39.685350514Z W1204 11:57:39.683958 1 reflector.go:484] sigs.k8s.io/controller-runtime/pkg/cache/internal/informers.go:106: watch of *v1.RoleBinding ended with: an error on the server ("unable to decode an event from the watch stream: context canceled") has prevented the request from succeeding 2025-12-04T11:57:39.685380285Z W1204 11:57:39.684655 1 reflector.go:484] sigs.k8s.io/controller-runtime/pkg/cache/internal/informers.go:106: watch of *v1.Role ended with: an error on the server ("unable to decode an event from the watch stream: context canceled") has prevented the request from succeeding 2025-12-04T11:57:39.685399956Z 2025-12-04T11:57:39.685Z INFO operator.init runtime/asm_amd64.s:1695 Wait completed, proceeding to shutdown the manager 2025-12-04T11:57:39.685440587Z W1204 11:57:39.684239 1 reflector.go:484] sigs.k8s.io/controller-runtime/pkg/cache/internal/informers.go:106: watch of *v1.Proxy ended with: an error on the server ("unable to decode an event from the watch stream: context canceled") has prevented the request from succeeding 2025-12-04T11:57:39.685471378Z W1204 11:57:39.684106 1 reflector.go:484] sigs.k8s.io/controller-runtime/pkg/cache/internal/informers.go:106: watch of *v1.Ingress ended with: an error on the server ("unable to decode an event from the watch stream: context canceled") has prevented the request from succeeding 2025-12-04T11:57:39.686278430Z 2025-12-04T11:57:39.685Z ERROR operator.main cobra/command.go:989 error starting {"error": "failed to wait for canary_controller caches to sync: timed out waiting for cache to be synced for Kind *v1.Route"} config/pod/openshift-ingress-operator/ingress-operator-8649c48786-cx2b2.json0000640000000000000000000002431315114274105025102 0ustar0000000000000000{"metadata":{"name":"ingress-operator-8649c48786-cx2b2","generateName":"ingress-operator-8649c48786-","namespace":"openshift-ingress-operator","uid":"b011b1f1-3235-4e20-825b-ce711c052407","resourceVersion":"11946","creationTimestamp":"2025-12-04T11:35:50Z","labels":{"name":"ingress-operator","pod-template-hash":"8649c48786"},"annotations":{"k8s.ovn.org/pod-networks":"{\"default\":{\"ip_addresses\":[\"10.128.0.20/23\"],\"mac_address\":\"0a:58:0a:80:00:14\",\"gateway_ips\":[\"10.128.0.1\"],\"routes\":[{\"dest\":\"10.128.0.0/16\",\"nextHop\":\"10.128.0.1\"},{\"dest\":\"172.30.0.0/16\",\"nextHop\":\"10.128.0.1\"},{\"dest\":\"169.254.0.5/32\",\"nextHop\":\"10.128.0.1\"},{\"dest\":\"100.64.0.0/16\",\"nextHop\":\"10.128.0.1\"}],\"ip_address\":\"10.128.0.20/23\",\"gateway_ip\":\"10.128.0.1\",\"role\":\"primary\"}}","k8s.v1.cni.cncf.io/network-status":"[{\n \"name\": \"ovn-kubernetes\",\n \"interface\": \"eth0\",\n \"ips\": [\n \"10.128.0.20\"\n ],\n \"mac\": \"0a:58:0a:80:00:14\",\n \"default\": true,\n \"dns\": {}\n}]","openshift.io/required-scc":"restricted-v2","openshift.io/scc":"restricted-v2","seccomp.security.alpha.kubernetes.io/pod":"runtime/default"},"ownerReferences":[{"apiVersion":"apps/v1","kind":"ReplicaSet","name":"ingress-operator-8649c48786","uid":"b8f6683f-8ce0-4bb9-bf4e-fb383622af51","controller":true,"blockOwnerDeletion":true}]},"spec":{"volumes":[{"name":"metrics-tls","secret":{"secretName":"metrics-tls","defaultMode":420}},{"name":"trusted-ca","configMap":{"name":"trusted-ca","items":[{"key":"ca-bundle.crt","path":"tls-ca-bundle.pem"}],"defaultMode":420}},{"name":"bound-sa-token","projected":{"sources":[{"serviceAccountToken":{"audience":"openshift","expirationSeconds":3600,"path":"token"}}],"defaultMode":420}},{"name":"kube-api-access-jq49f","projected":{"sources":[{"serviceAccountToken":{"expirationSeconds":3607,"path":"token"}},{"configMap":{"name":"kube-root-ca.crt","items":[{"key":"ca.crt","path":"ca.crt"}]}},{"downwardAPI":{"items":[{"path":"namespace","fieldRef":{"apiVersion":"v1","fieldPath":"metadata.namespace"}}]}},{"configMap":{"name":"openshift-service-ca.crt","items":[{"key":"service-ca.crt","path":"service-ca.crt"}]}}],"defaultMode":420}}],"containers":[{"name":"ingress-operator","image":"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:831f30660844091d6154e2674d3a9da6f34271bf8a2c40b56f7416066318742b","command":["ingress-operator","start","--namespace","$(WATCH_NAMESPACE)","--image","$(IMAGE)","--canary-image","$(CANARY_IMAGE)","--release-version","$(RELEASE_VERSION)"],"env":[{"name":"RELEASE_VERSION","value":"4.18.29"},{"name":"WATCH_NAMESPACE","valueFrom":{"fieldRef":{"apiVersion":"v1","fieldPath":"metadata.namespace"}}},{"name":"IMAGE","value":"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:2b3d313c599852b3543ee5c3a62691bd2d1bbad12c2e1c610cd71a1dec6eea32"},{"name":"CANARY_IMAGE","value":"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:831f30660844091d6154e2674d3a9da6f34271bf8a2c40b56f7416066318742b"}],"resources":{"requests":{"cpu":"10m","memory":"56Mi"}},"volumeMounts":[{"name":"trusted-ca","readOnly":true,"mountPath":"/etc/pki/ca-trust/extracted/pem"},{"name":"bound-sa-token","readOnly":true,"mountPath":"/var/run/secrets/openshift/serviceaccount"},{"name":"kube-api-access-jq49f","readOnly":true,"mountPath":"/var/run/secrets/kubernetes.io/serviceaccount"}],"terminationMessagePath":"/dev/termination-log","terminationMessagePolicy":"FallbackToLogsOnError","imagePullPolicy":"IfNotPresent","securityContext":{"capabilities":{"drop":["ALL"]},"runAsUser":1000140000,"allowPrivilegeEscalation":false}},{"name":"kube-rbac-proxy","image":"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69ffd8f8dcceedc2d6eb306cea33f8beabc1be1308cd5f4ee8b9a8e3eab9843","args":["--logtostderr","--secure-listen-address=:9393","--tls-cipher-suites=TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_RSA_WITH_AES_128_CBC_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256,TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256","--upstream=http://127.0.0.1:60000/","--tls-cert-file=/etc/tls/private/tls.crt","--tls-private-key-file=/etc/tls/private/tls.key"],"ports":[{"name":"metrics","containerPort":9393,"protocol":"TCP"}],"resources":{"requests":{"cpu":"10m","memory":"40Mi"}},"volumeMounts":[{"name":"metrics-tls","readOnly":true,"mountPath":"/etc/tls/private"},{"name":"kube-api-access-jq49f","readOnly":true,"mountPath":"/var/run/secrets/kubernetes.io/serviceaccount"}],"terminationMessagePath":"/dev/termination-log","terminationMessagePolicy":"FallbackToLogsOnError","imagePullPolicy":"IfNotPresent","securityContext":{"capabilities":{"drop":["ALL"]},"runAsUser":1000140000,"allowPrivilegeEscalation":false}}],"restartPolicy":"Always","terminationGracePeriodSeconds":30,"dnsPolicy":"ClusterFirst","nodeSelector":{"kubernetes.io/os":"linux","node-role.kubernetes.io/master":""},"serviceAccountName":"ingress-operator","serviceAccount":"ingress-operator","nodeName":"master-0","securityContext":{"seLinuxOptions":{"level":"s0:c12,c4"},"runAsNonRoot":true,"fsGroup":1000140000,"seccompProfile":{"type":"RuntimeDefault"}},"schedulerName":"default-scheduler","tolerations":[{"key":"node-role.kubernetes.io/master","operator":"Exists","effect":"NoSchedule"},{"key":"node.kubernetes.io/unreachable","operator":"Exists","effect":"NoExecute","tolerationSeconds":120},{"key":"node.kubernetes.io/not-ready","operator":"Exists","effect":"NoExecute","tolerationSeconds":120},{"key":"node.kubernetes.io/memory-pressure","operator":"Exists","effect":"NoSchedule"}],"priorityClassName":"system-cluster-critical","priority":2000000000,"enableServiceLinks":true,"preemptionPolicy":"PreemptLowerPriority"},"status":{"phase":"Running","conditions":[{"type":"PodReadyToStartContainers","status":"True","lastProbeTime":null,"lastTransitionTime":"2025-12-04T11:38:20Z"},{"type":"Initialized","status":"True","lastProbeTime":null,"lastTransitionTime":"2025-12-04T11:37:30Z"},{"type":"Ready","status":"True","lastProbeTime":null,"lastTransitionTime":"2025-12-04T11:59:31Z"},{"type":"ContainersReady","status":"True","lastProbeTime":null,"lastTransitionTime":"2025-12-04T11:59:31Z"},{"type":"PodScheduled","status":"True","lastProbeTime":null,"lastTransitionTime":"2025-12-04T11:37:30Z"}],"hostIP":"192.168.32.10","hostIPs":[{"ip":"192.168.32.10"}],"podIP":"10.128.0.20","podIPs":[{"ip":"10.128.0.20"}],"startTime":"2025-12-04T11:37:30Z","containerStatuses":[{"name":"ingress-operator","state":{"running":{"startedAt":"2025-12-04T11:59:30Z"}},"lastState":{"terminated":{"exitCode":1,"reason":"Error","message":"ller ended with: an error on the server (\"unable to decode an event from the watch stream: context canceled\") has prevented the request from succeeding\n2025-12-04T11:57:39.685Z\tINFO\toperator.init.controller-runtime.metrics\truntime/asm_amd64.s:1695\tShutting down metrics server with timeout of 1 minute\nW1204 11:57:39.683922 1 reflector.go:484] sigs.k8s.io/controller-runtime/pkg/cache/internal/informers.go:106: watch of *v1.Service ended with: an error on the server (\"unable to decode an event from the watch stream: context canceled\") has prevented the request from succeeding\nW1204 11:57:39.683958 1 reflector.go:484] sigs.k8s.io/controller-runtime/pkg/cache/internal/informers.go:106: watch of *v1.RoleBinding ended with: an error on the server (\"unable to decode an event from the watch stream: context canceled\") has prevented the request from succeeding\nW1204 11:57:39.684655 1 reflector.go:484] sigs.k8s.io/controller-runtime/pkg/cache/internal/informers.go:106: watch of *v1.Role ended with: an error on the server (\"unable to decode an event from the watch stream: context canceled\") has prevented the request from succeeding\n2025-12-04T11:57:39.685Z\tINFO\toperator.init\truntime/asm_amd64.s:1695\tWait completed, proceeding to shutdown the manager\nW1204 11:57:39.684239 1 reflector.go:484] sigs.k8s.io/controller-runtime/pkg/cache/internal/informers.go:106: watch of *v1.Proxy ended with: an error on the server (\"unable to decode an event from the watch stream: context canceled\") has prevented the request from succeeding\nW1204 11:57:39.684106 1 reflector.go:484] sigs.k8s.io/controller-runtime/pkg/cache/internal/informers.go:106: watch of *v1.Ingress ended with: an error on the server (\"unable to decode an event from the watch stream: context canceled\") has prevented the request from succeeding\n2025-12-04T11:57:39.685Z\tERROR\toperator.main\tcobra/command.go:989\terror starting\t{\"error\": \"failed to wait for canary_controller caches to sync: timed out waiting for cache to be synced for Kind *v1.Route\"}\n","startedAt":"2025-12-04T11:54:38Z","finishedAt":"2025-12-04T11:57:39Z","containerID":"cri-o://6b81ee1ae28f16389ded7b634a062f62bd333e247fbc631f5edd514e5d0827da"}},"ready":true,"restartCount":6,"image":"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:831f30660844091d6154e2674d3a9da6f34271bf8a2c40b56f7416066318742b","imageID":"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:831f30660844091d6154e2674d3a9da6f34271bf8a2c40b56f7416066318742b","containerID":"cri-o://6e2b9bac78892a0ed4bd84f11f2eb0c1268bd4937b6948a7c059ed07470dce0e","started":true,"volumeMounts":[{"name":"trusted-ca","mountPath":"/etc/pki/ca-trust/extracted/pem","readOnly":true,"recursiveReadOnly":"Disabled"},{"name":"bound-sa-token","mountPath":"/var/run/secrets/openshift/serviceaccount","readOnly":true,"recursiveReadOnly":"Disabled"},{"name":"kube-api-access-jq49f","mountPath":"/var/run/secrets/kubernetes.io/serviceaccount","readOnly":true,"recursiveReadOnly":"Disabled"}]},{"name":"kube-rbac-proxy","state":{"running":{"startedAt":"2025-12-04T11:38:19Z"}},"lastState":{},"ready":true,"restartCount":0,"image":"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69ffd8f8dcceedc2d6eb306cea33f8beabc1be1308cd5f4ee8b9a8e3eab9843","imageID":"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69ffd8f8dcceedc2d6eb306cea33f8beabc1be1308cd5f4ee8b9a8e3eab9843","containerID":"cri-o://51cef92d704724c90de3ae256673cce5ec9954aeee7c8eb66419deadfabecfec","started":true,"volumeMounts":[{"name":"metrics-tls","mountPath":"/etc/tls/private","readOnly":true,"recursiveReadOnly":"Disabled"},{"name":"kube-api-access-jq49f","mountPath":"/var/run/secrets/kubernetes.io/serviceaccount","readOnly":true,"recursiveReadOnly":"Disabled"}]}],"qosClass":"Burstable"}}cluster-image-registry-operator_current.log0000640000000000000000000007432315114274105040522 0ustar0000000000000000config/pod/openshift-image-registry/logs/cluster-image-registry-operator-6fb9f88b7-tgvfl2025-12-04T11:59:53.676996896Z E1204 11:59:53.676988 1 imageregistrycertificates.go:144] ImageRegistryCertificatesController: unable to sync: config.imageregistry.operator.openshift.io "cluster" not found, requeuing 2025-12-04T11:59:54.146261148Z E1204 11:59:54.146069 1 clusteroperator.go:104] unable to sync ClusterOperatorStatusController: config.imageregistry.operator.openshift.io "cluster" not found, requeuing 2025-12-04T11:59:54.321867730Z E1204 11:59:54.321740 1 clusteroperator.go:104] unable to sync ClusterOperatorStatusController: config.imageregistry.operator.openshift.io "cluster" not found, requeuing 2025-12-04T11:59:54.733509067Z I1204 11:59:54.733398 1 caconfig.go:127] unable to get the service name to add service-ca.crt 2025-12-04T11:59:54.733509067Z I1204 11:59:54.733471 1 imageregistryca.go:130] unable to get the service name to add service-ca.crt 2025-12-04T11:59:54.733592670Z E1204 11:59:54.733502 1 imageregistrycertificates.go:144] ImageRegistryCertificatesController: unable to sync: config.imageregistry.operator.openshift.io "cluster" not found, requeuing 2025-12-04T11:59:55.122872688Z E1204 11:59:55.122738 1 clusteroperator.go:104] unable to sync ClusterOperatorStatusController: config.imageregistry.operator.openshift.io "cluster" not found, requeuing 2025-12-04T11:59:56.846787907Z E1204 11:59:56.846674 1 clusteroperator.go:104] unable to sync ClusterOperatorStatusController: config.imageregistry.operator.openshift.io "cluster" not found, requeuing 2025-12-04T11:59:57.813010291Z E1204 11:59:57.812920 1 clusteroperator.go:104] unable to sync ClusterOperatorStatusController: config.imageregistry.operator.openshift.io "cluster" not found, requeuing 2025-12-04T11:59:57.813211506Z E1204 11:59:57.813157 1 clusteroperator.go:104] unable to sync ClusterOperatorStatusController: config.imageregistry.operator.openshift.io "cluster" not found, requeuing 2025-12-04T11:59:57.865481424Z E1204 11:59:57.864135 1 clusteroperator.go:104] unable to sync ClusterOperatorStatusController: config.imageregistry.operator.openshift.io "cluster" not found, requeuing 2025-12-04T11:59:58.020532410Z E1204 11:59:58.020413 1 clusteroperator.go:104] unable to sync ClusterOperatorStatusController: config.imageregistry.operator.openshift.io "cluster" not found, requeuing 2025-12-04T11:59:58.047793134Z E1204 11:59:58.047032 1 clusteroperator.go:104] unable to sync ClusterOperatorStatusController: config.imageregistry.operator.openshift.io "cluster" not found, requeuing 2025-12-04T11:59:58.789377261Z I1204 11:59:58.789290 1 caconfig.go:127] unable to get the service name to add service-ca.crt 2025-12-04T11:59:58.789430472Z I1204 11:59:58.789395 1 imageregistryca.go:130] unable to get the service name to add service-ca.crt 2025-12-04T11:59:58.789458043Z E1204 11:59:58.789439 1 imageregistrycertificates.go:144] ImageRegistryCertificatesController: unable to sync: config.imageregistry.operator.openshift.io "cluster" not found, requeuing 2025-12-04T11:59:58.789488554Z I1204 11:59:58.789463 1 caconfig.go:127] unable to get the service name to add service-ca.crt 2025-12-04T11:59:58.789488554Z I1204 11:59:58.789481 1 imageregistryca.go:130] unable to get the service name to add service-ca.crt 2025-12-04T11:59:58.790149555Z E1204 11:59:58.789495 1 imageregistrycertificates.go:144] ImageRegistryCertificatesController: unable to sync: config.imageregistry.operator.openshift.io "cluster" not found, requeuing 2025-12-04T12:00:01.482178655Z I1204 12:00:01.482060 1 imageconfig.go:112] Started ImageConfigController 2025-12-04T12:00:01.482178655Z I1204 12:00:01.482132 1 controller.go:454] Starting Controller 2025-12-04T12:00:01.482278988Z I1204 12:00:01.482251 1 bootstrap.go:39] generating registry custom resource 2025-12-04T12:00:01.487156012Z E1204 12:00:01.487081 1 imageconfig.go:132] ImageConfigController: unable to sync: config.imageregistry.operator.openshift.io "cluster" not found, requeuing 2025-12-04T12:00:01.790343505Z I1204 12:00:01.790254 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go/informers/factory.go:160 2025-12-04T12:00:01.790482719Z I1204 12:00:01.790441 1 caconfig.go:127] unable to get the service name to add service-ca.crt 2025-12-04T12:00:01.790538951Z I1204 12:00:01.790509 1 imageregistryca.go:130] unable to get the service name to add service-ca.crt 2025-12-04T12:00:01.845481811Z I1204 12:00:01.845409 1 generator.go:63] object *v1.ClusterOperator, Name=image-registry updated: removed:apiVersion="config.openshift.io/v1", removed:kind="ClusterOperator", added:metadata.managedFields.2.apiVersion="config.openshift.io/v1", added:metadata.managedFields.2.fieldsType="FieldsV1", added:metadata.managedFields.2.manager="cluster-image-registry-operator", added:metadata.managedFields.2.operation="Update", added:metadata.managedFields.2.subresource="status", added:metadata.managedFields.2.time="2025-12-04T12:00:01Z", changed:metadata.resourceVersion={"747" -> "12177"}, added:status.conditions.0.lastTransitionTime="2025-12-04T12:00:01Z", added:status.conditions.0.message="ImagePrunerAvailable: Pruner CronJob has been created", added:status.conditions.0.reason="AsExpected", added:status.conditions.0.status="True", added:status.conditions.0.type="Available", added:status.conditions.1.lastTransitionTime="2025-12-04T12:00:01Z", added:status.conditions.1.reason="AsExpected", added:status.conditions.1.status="False", added:status.conditions.1.type="Progressing", added:status.conditions.2.lastTransitionTime="2025-12-04T12:00:01Z", added:status.conditions.2.reason="AsExpected", added:status.conditions.2.status="False", added:status.conditions.2.type="Degraded", added:status.relatedObjects.0.group="imageregistry.operator.openshift.io", added:status.relatedObjects.0.name="cluster", added:status.relatedObjects.0.resource="configs", added:status.relatedObjects.1.group="imageregistry.operator.openshift.io", added:status.relatedObjects.1.name="cluster", added:status.relatedObjects.1.resource="imagepruners", added:status.relatedObjects.2.group="rbac.authorization.k8s.io", added:status.relatedObjects.2.name="system:registry", added:status.relatedObjects.2.resource="clusterroles", added:status.relatedObjects.3.group="rbac.authorization.k8s.io", added:status.relatedObjects.3.name="registry-registry-role", added:status.relatedObjects.3.resource="clusterrolebindings", added:status.relatedObjects.4.group="rbac.authorization.k8s.io", added:status.relatedObjects.4.name="openshift-image-registry-pruner", added:status.relatedObjects.4.resource="clusterrolebindings", added:status.relatedObjects.5.group="", added:status.relatedObjects.5.name="openshift-image-registry", added:status.relatedObjects.5.resource="namespaces", added:status.versions.0.name="operator", added:status.versions.0.version="4.18.29" 2025-12-04T12:00:02.741942072Z I1204 12:00:02.741886 1 controller.go:340] object changed: *v1.Config, Name=cluster (status=true): added:status.conditions.0.lastTransitionTime="2025-12-04T12:00:02Z", added:status.conditions.0.message="All registry resources are removed", added:status.conditions.0.reason="Removed", added:status.conditions.0.status="False", added:status.conditions.0.type="Progressing", added:status.conditions.1.lastTransitionTime="2025-12-04T12:00:02Z", added:status.conditions.1.message="The registry is removed", added:status.conditions.1.reason="Removed", added:status.conditions.1.status="True", added:status.conditions.1.type="Available", added:status.conditions.2.lastTransitionTime="2025-12-04T12:00:02Z", added:status.conditions.2.message="The registry is removed", added:status.conditions.2.reason="Removed", added:status.conditions.2.status="False", added:status.conditions.2.type="Degraded", added:status.conditions.3.lastTransitionTime="2025-12-04T12:00:02Z", added:status.conditions.3.message="The registry is removed", added:status.conditions.3.reason="Removed", added:status.conditions.3.status="True", added:status.conditions.3.type="Removed", added:status.observedGeneration="1.000000" 2025-12-04T12:00:02.744054659Z I1204 12:00:02.742833 1 caconfig.go:127] unable to get the service name to add service-ca.crt 2025-12-04T12:00:02.744054659Z I1204 12:00:02.742910 1 imageregistryca.go:130] unable to get the service name to add service-ca.crt 2025-12-04T12:00:03.470939351Z E1204 12:00:03.469197 1 controller.go:379] unable to sync: Operation cannot be fulfilled on configs.imageregistry.operator.openshift.io "cluster": the object has been modified; please apply your changes to the latest version and try again, requeuing 2025-12-04T12:00:03.865715014Z I1204 12:00:03.865650 1 controller.go:340] object changed: *v1.Config, Name=cluster (status=true): added:status.conditions.3.lastTransitionTime="2025-12-04T12:00:03Z", added:status.conditions.3.message="All registry resources are removed", added:status.conditions.3.reason="Removed", added:status.conditions.3.status="False", added:status.conditions.3.type="Progressing", added:status.conditions.4.lastTransitionTime="2025-12-04T12:00:03Z", added:status.conditions.4.message="The registry is removed", added:status.conditions.4.reason="Removed", added:status.conditions.4.status="True", added:status.conditions.4.type="Available", added:status.conditions.5.lastTransitionTime="2025-12-04T12:00:03Z", added:status.conditions.5.message="The registry is removed", added:status.conditions.5.reason="Removed", added:status.conditions.5.status="False", added:status.conditions.5.type="Degraded", added:status.conditions.6.lastTransitionTime="2025-12-04T12:00:03Z", added:status.conditions.6.message="The registry is removed", added:status.conditions.6.reason="Removed", added:status.conditions.6.status="True", added:status.conditions.6.type="Removed", added:status.observedGeneration="1.000000" 2025-12-04T12:00:04.248947711Z I1204 12:00:04.248869 1 generator.go:63] object *v1.ClusterOperator, Name=image-registry updated: changed:metadata.managedFields.2.time={"2025-12-04T12:00:01Z" -> "2025-12-04T12:00:04Z"}, changed:metadata.resourceVersion={"12177" -> "12192"}, changed:status.conditions.0.message={"ImagePrunerAvailable: Pruner CronJob has been created" -> "Available: The registry is removed\nImagePrunerAvailable: Pruner CronJob has been created"}, changed:status.conditions.0.reason={"AsExpected" -> "Removed"}, added:status.conditions.1.message="Progressing: All registry resources are removed", changed:status.conditions.1.reason={"AsExpected" -> "Removed"}, added:status.conditions.2.message="Degraded: The registry is removed", changed:status.conditions.2.reason={"AsExpected" -> "Removed"} 2025-12-04T12:00:04.301151224Z I1204 12:00:04.300970 1 controller.go:340] object changed: *v1.Config, Name=cluster (status=true): added:status.conditions.3.lastTransitionTime="2025-12-04T12:00:04Z", added:status.conditions.3.message="All registry resources are removed", added:status.conditions.3.reason="Removed", added:status.conditions.3.status="False", added:status.conditions.3.type="Progressing", added:status.conditions.4.lastTransitionTime="2025-12-04T12:00:04Z", added:status.conditions.4.message="The registry is removed", added:status.conditions.4.reason="Removed", added:status.conditions.4.status="True", added:status.conditions.4.type="Available", added:status.conditions.5.lastTransitionTime="2025-12-04T12:00:04Z", added:status.conditions.5.message="The registry is removed", added:status.conditions.5.reason="Removed", added:status.conditions.5.status="False", added:status.conditions.5.type="Degraded", added:status.conditions.6.lastTransitionTime="2025-12-04T12:00:04Z", added:status.conditions.6.message="The registry is removed", added:status.conditions.6.reason="Removed", added:status.conditions.6.status="True", added:status.conditions.6.type="Removed", added:status.observedGeneration="1.000000" 2025-12-04T12:00:04.306324278Z E1204 12:00:04.306236 1 controller.go:379] unable to sync: Operation cannot be fulfilled on configs.imageregistry.operator.openshift.io "cluster": the object has been modified; please apply your changes to the latest version and try again, requeuing 2025-12-04T12:00:04.485542514Z I1204 12:00:04.485239 1 controller.go:340] object changed: *v1.Config, Name=cluster (status=true): changed:status.conditions.3.lastTransitionTime={"2025-12-04T12:00:03Z" -> "2025-12-04T12:00:04Z"} 2025-12-04T12:00:04.801912553Z I1204 12:00:04.799315 1 caconfig.go:127] unable to get the service name to add service-ca.crt 2025-12-04T12:00:04.801912553Z I1204 12:00:04.799396 1 imageregistryca.go:130] unable to get the service name to add service-ca.crt 2025-12-04T12:00:04.852407502Z I1204 12:00:04.850431 1 caconfig.go:127] unable to get the service name to add service-ca.crt 2025-12-04T12:00:04.852407502Z I1204 12:00:04.850478 1 imageregistryca.go:130] unable to get the service name to add service-ca.crt 2025-12-04T12:00:05.120436391Z I1204 12:00:05.120368 1 controller.go:340] object changed: *v1.Config, Name=cluster (status=true): changed:status.conditions.3.lastTransitionTime={"2025-12-04T12:00:04Z" -> "2025-12-04T12:00:05Z"} 2025-12-04T12:00:05.997893841Z I1204 12:00:05.997799 1 controller.go:340] object changed: *v1.Config, Name=cluster (status=true): 2025-12-04T12:00:10.617876762Z I1204 12:00:10.614204 1 caconfig.go:127] unable to get the service name to add service-ca.crt 2025-12-04T12:00:10.617876762Z I1204 12:00:10.614273 1 imageregistryca.go:130] unable to get the service name to add service-ca.crt 2025-12-04T12:00:10.640890550Z I1204 12:00:10.640361 1 caconfig.go:127] unable to get the service name to add service-ca.crt 2025-12-04T12:00:10.640890550Z I1204 12:00:10.640431 1 imageregistryca.go:130] unable to get the service name to add service-ca.crt 2025-12-04T12:00:10.681509067Z I1204 12:00:10.680884 1 controller.go:340] object changed: *v1.Config, Name=cluster (status=true): changed:status.conditions.3.lastTransitionTime={"2025-12-04T12:00:05Z" -> "2025-12-04T12:00:10Z"} 2025-12-04T12:00:10.721688790Z I1204 12:00:10.721589 1 controller.go:340] object changed: *v1.Config, Name=cluster (status=true): 2025-12-04T12:00:11.010181897Z I1204 12:00:11.010074 1 caconfig.go:127] unable to get the service name to add service-ca.crt 2025-12-04T12:00:11.010181897Z I1204 12:00:11.010159 1 imageregistryca.go:130] unable to get the service name to add service-ca.crt 2025-12-04T12:00:11.020723570Z I1204 12:00:11.020617 1 caconfig.go:127] unable to get the service name to add service-ca.crt 2025-12-04T12:00:11.020723570Z I1204 12:00:11.020688 1 imageregistryca.go:130] unable to get the service name to add service-ca.crt 2025-12-04T12:00:11.056035598Z I1204 12:00:11.055644 1 controller.go:340] object changed: *v1.Config, Name=cluster (status=true): changed:status.conditions.3.lastTransitionTime={"2025-12-04T12:00:10Z" -> "2025-12-04T12:00:11Z"} 2025-12-04T12:00:11.865471354Z I1204 12:00:11.865094 1 controller.go:340] object changed: *v1.Config, Name=cluster (status=true): changed:status.conditions.3.lastTransitionTime={"2025-12-04T12:00:10Z" -> "2025-12-04T12:00:11Z"} 2025-12-04T12:00:11.878153586Z E1204 12:00:11.877324 1 controller.go:379] unable to sync: Operation cannot be fulfilled on configs.imageregistry.operator.openshift.io "cluster": the object has been modified; please apply your changes to the latest version and try again, requeuing 2025-12-04T12:00:12.656383314Z I1204 12:00:12.656321 1 controller.go:340] object changed: *v1.Config, Name=cluster (status=true): changed:status.conditions.3.lastTransitionTime={"2025-12-04T12:00:11Z" -> "2025-12-04T12:00:12Z"} 2025-12-04T12:00:13.442612784Z I1204 12:00:13.442419 1 caconfig.go:127] unable to get the service name to add service-ca.crt 2025-12-04T12:00:13.442612784Z I1204 12:00:13.442525 1 imageregistryca.go:130] unable to get the service name to add service-ca.crt 2025-12-04T12:00:13.454586213Z I1204 12:00:13.454471 1 caconfig.go:127] unable to get the service name to add service-ca.crt 2025-12-04T12:00:13.454631214Z I1204 12:00:13.454592 1 imageregistryca.go:130] unable to get the service name to add service-ca.crt 2025-12-04T12:00:13.459913733Z I1204 12:00:13.458425 1 controller.go:340] object changed: *v1.Config, Name=cluster (status=true): changed:status.conditions.3.lastTransitionTime={"2025-12-04T12:00:12Z" -> "2025-12-04T12:00:13Z"} 2025-12-04T12:00:14.253322450Z I1204 12:00:14.253241 1 controller.go:340] object changed: *v1.Config, Name=cluster (status=true): changed:status.conditions.3.lastTransitionTime={"2025-12-04T12:00:12Z" -> "2025-12-04T12:00:14Z"} 2025-12-04T12:00:14.258339129Z E1204 12:00:14.258262 1 controller.go:379] unable to sync: Operation cannot be fulfilled on configs.imageregistry.operator.openshift.io "cluster": the object has been modified; please apply your changes to the latest version and try again, requeuing 2025-12-04T12:00:15.071282436Z I1204 12:00:15.069569 1 controller.go:340] object changed: *v1.Config, Name=cluster (status=true): changed:status.conditions.3.lastTransitionTime={"2025-12-04T12:00:13Z" -> "2025-12-04T12:00:15Z"} 2025-12-04T12:00:15.858187108Z I1204 12:00:15.858084 1 controller.go:340] object changed: *v1.Config, Name=cluster (status=true): changed:status.conditions.3.lastTransitionTime={"2025-12-04T12:00:13Z" -> "2025-12-04T12:00:15Z"} 2025-12-04T12:00:15.864041434Z E1204 12:00:15.863953 1 controller.go:379] unable to sync: Operation cannot be fulfilled on configs.imageregistry.operator.openshift.io "cluster": the object has been modified; please apply your changes to the latest version and try again, requeuing 2025-12-04T12:00:16.670280648Z I1204 12:00:16.670209 1 controller.go:340] object changed: *v1.Config, Name=cluster (status=true): changed:status.conditions.3.lastTransitionTime={"2025-12-04T12:00:15Z" -> "2025-12-04T12:00:16Z"} 2025-12-04T12:00:16.848709229Z I1204 12:00:16.848579 1 caconfig.go:127] unable to get the service name to add service-ca.crt 2025-12-04T12:00:16.848709229Z I1204 12:00:16.848677 1 imageregistryca.go:130] unable to get the service name to add service-ca.crt 2025-12-04T12:00:16.858731686Z I1204 12:00:16.858033 1 caconfig.go:127] unable to get the service name to add service-ca.crt 2025-12-04T12:00:16.858731686Z I1204 12:00:16.858116 1 imageregistryca.go:130] unable to get the service name to add service-ca.crt 2025-12-04T12:00:17.453053679Z I1204 12:00:17.452919 1 controller.go:340] object changed: *v1.Config, Name=cluster (status=true): changed:status.conditions.3.lastTransitionTime={"2025-12-04T12:00:15Z" -> "2025-12-04T12:00:17Z"} 2025-12-04T12:00:17.457180170Z E1204 12:00:17.457139 1 controller.go:379] unable to sync: Operation cannot be fulfilled on configs.imageregistry.operator.openshift.io "cluster": the object has been modified; please apply your changes to the latest version and try again, requeuing 2025-12-04T12:00:18.256743643Z I1204 12:00:18.256651 1 controller.go:340] object changed: *v1.Config, Name=cluster (status=true): changed:status.conditions.3.lastTransitionTime={"2025-12-04T12:00:16Z" -> "2025-12-04T12:00:18Z"} 2025-12-04T12:00:19.054019184Z I1204 12:00:19.053933 1 controller.go:340] object changed: *v1.Config, Name=cluster (status=true): changed:status.conditions.3.lastTransitionTime={"2025-12-04T12:00:16Z" -> "2025-12-04T12:00:19Z"} 2025-12-04T12:00:19.059036423Z E1204 12:00:19.058976 1 controller.go:379] unable to sync: Operation cannot be fulfilled on configs.imageregistry.operator.openshift.io "cluster": the object has been modified; please apply your changes to the latest version and try again, requeuing 2025-12-04T12:00:19.855955302Z I1204 12:00:19.855801 1 controller.go:340] object changed: *v1.Config, Name=cluster (status=true): changed:status.conditions.3.lastTransitionTime={"2025-12-04T12:00:18Z" -> "2025-12-04T12:00:19Z"} 2025-12-04T12:00:20.129459724Z I1204 12:00:20.129274 1 caconfig.go:127] unable to get the service name to add service-ca.crt 2025-12-04T12:00:20.129459724Z I1204 12:00:20.129423 1 imageregistryca.go:130] unable to get the service name to add service-ca.crt 2025-12-04T12:00:20.136311862Z I1204 12:00:20.136075 1 caconfig.go:127] unable to get the service name to add service-ca.crt 2025-12-04T12:00:20.136311862Z I1204 12:00:20.136165 1 imageregistryca.go:130] unable to get the service name to add service-ca.crt 2025-12-04T12:00:20.660104851Z I1204 12:00:20.660015 1 controller.go:340] object changed: *v1.Config, Name=cluster (status=true): changed:status.conditions.3.lastTransitionTime={"2025-12-04T12:00:19Z" -> "2025-12-04T12:00:20Z"} 2025-12-04T12:00:21.614208607Z I1204 12:00:21.614118 1 controller.go:340] object changed: *v1.Config, Name=cluster (status=true): changed:status.conditions.3.lastTransitionTime={"2025-12-04T12:00:20Z" -> "2025-12-04T12:00:21Z"} 2025-12-04T12:00:22.475399151Z I1204 12:00:22.475325 1 controller.go:340] object changed: *v1.Config, Name=cluster (status=true): changed:status.conditions.3.lastTransitionTime={"2025-12-04T12:00:21Z" -> "2025-12-04T12:00:22Z"} 2025-12-04T12:00:23.059011705Z I1204 12:00:23.058924 1 controller.go:340] object changed: *v1.Config, Name=cluster (status=true): changed:status.conditions.3.lastTransitionTime={"2025-12-04T12:00:22Z" -> "2025-12-04T12:00:23Z"} 2025-12-04T12:00:23.877820458Z I1204 12:00:23.877706 1 controller.go:340] object changed: *v1.Config, Name=cluster (status=true): 2025-12-04T12:01:39.772829336Z I1204 12:01:39.772720 1 caconfig.go:127] unable to get the service name to add service-ca.crt 2025-12-04T12:01:39.772891368Z I1204 12:01:39.772831 1 imageregistryca.go:130] unable to get the service name to add service-ca.crt 2025-12-04T12:01:39.779189687Z I1204 12:01:39.779118 1 nodecadaemon.go:100] NodeCADaemonController processing requeued item changes 2025-12-04T12:01:39.806523993Z I1204 12:01:39.804415 1 event.go:377] Event(v1.ObjectReference{Kind:"Deployment", Namespace:"openshift-image-registry", Name:"cluster-image-registry-operator", UID:"19f08704-38bb-4621-9ec0-73544763138a", APIVersion:"apps/v1", ResourceVersion:"", FieldPath:""}): type: 'Normal' reason: 'DaemonSetCreated' Created DaemonSet.apps/node-ca -n openshift-image-registry because it was missing 2025-12-04T12:01:39.817040725Z I1204 12:01:39.816951 1 generator.go:46] object *v1.DaemonSet, Namespace=openshift-image-registry, Name=node-ca created: metadata.annotations.deprecated.daemonset.template.generation="1", metadata.annotations.operator.openshift.io/spec-hash="8d619f6b42bf25f6e25d42b4e5a389a0a1d382dd737741e9100884612d9bafc9", metadata.creationTimestamp="2025-12-04T12:01:39Z", metadata.generation="1.000000", metadata.managedFields.0.apiVersion="apps/v1", metadata.managedFields.0.fieldsType="FieldsV1", metadata.managedFields.0.manager="cluster-image-registry-operator", metadata.managedFields.0.operation="Update", metadata.managedFields.0.time="2025-12-04T12:01:39Z", metadata.name="node-ca", metadata.namespace="openshift-image-registry", metadata.resourceVersion="14283", metadata.uid="aa466c53-ae80-470a-98f9-64783b5ab7d5", spec.revisionHistoryLimit="10.000000", spec.selector.matchLabels.name="node-ca", spec.template.metadata.annotations.openshift.io/required-scc="privileged", spec.template.metadata.annotations.target.workload.openshift.io/management="{\"effect\": \"PreferredDuringScheduling\"}", spec.template.metadata.creationTimestamp="nil", spec.template.metadata.labels.name="node-ca", spec.template.spec.containers.0.command.0="/bin/sh", spec.template.spec.containers.0.command.1="-c", spec.template.spec.containers.0.command.2="trap 'jobs -p | xargs -r kill; echo shutting down node-ca; exit 0' TERM\nwhile [ true ];\ndo\n for f in $(ls /tmp/serviceca); do\n echo $f\n ca_file_path=\"/tmp/serviceca/${f}\"\n f=$(echo $f | sed -r 's/(.*)\\.\\./\\1:/')\n reg_dir_path=\"/etc/docker/certs.d/${f}\"\n if [ -e \"${reg_dir_path}\" ]; then\n cp -u $ca_file_path $reg_dir_path/ca.crt\n else\n mkdir $reg_dir_path\n cp $ca_file_path $reg_dir_path/ca.crt\n fi\n done\n for d in $(ls /etc/docker/certs.d); do\n echo $d\n dp=$(echo $d | sed -r 's/(.*):/\\1\\.\\./')\n reg_conf_path=\"/tmp/serviceca/${dp}\"\n if [ ! -e \"${reg_conf_path}\" ]; then\n rm -rf /etc/docker/certs.d/$d\n fi\n done\n sleep 60 & wait ${!}\ndone\n", spec.template.spec.containers.0.image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3ebe19b23694155a15d0968968fdee3dcf200ab9718ae1fcbd05f4d24960b827", spec.template.spec.containers.0.imagePullPolicy="IfNotPresent", spec.template.spec.containers.0.name="node-ca", spec.template.spec.containers.0.resources.requests.cpu="10m", spec.template.spec.containers.0.resources.requests.memory="10Mi", spec.template.spec.containers.0.securityContext.privileged="true", spec.template.spec.containers.0.securityContext.runAsGroup="0.000000", spec.template.spec.containers.0.securityContext.runAsUser="1001.000000", spec.template.spec.containers.0.terminationMessagePath="/dev/termination-log", spec.template.spec.containers.0.terminationMessagePolicy="FallbackToLogsOnError", spec.template.spec.containers.0.volumeMounts.0.mountPath="/tmp/serviceca", spec.template.spec.containers.0.volumeMounts.0.name="serviceca", spec.template.spec.containers.0.volumeMounts.1.mountPath="/etc/docker/certs.d", spec.template.spec.containers.0.volumeMounts.1.name="host", spec.template.spec.dnsPolicy="ClusterFirst", spec.template.spec.hostNetwork="true", spec.template.spec.nodeSelector.kubernetes.io/os="linux", spec.template.spec.priorityClassName="system-cluster-critical", spec.template.spec.restartPolicy="Always", spec.template.spec.schedulerName="default-scheduler", spec.template.spec.serviceAccount="node-ca", spec.template.spec.serviceAccountName="node-ca", spec.template.spec.terminationGracePeriodSeconds="30.000000", spec.template.spec.tolerations.0.operator="Exists", spec.template.spec.volumes.0.hostPath.path="/etc/docker/certs.d", spec.template.spec.volumes.0.hostPath.type="", spec.template.spec.volumes.0.name="host", spec.template.spec.volumes.1.configMap.defaultMode="420.000000", spec.template.spec.volumes.1.configMap.name="image-registry-certificates", spec.template.spec.volumes.1.name="serviceca", spec.updateStrategy.rollingUpdate.maxSurge="0.000000", spec.updateStrategy.rollingUpdate.maxUnavailable="10%", spec.updateStrategy.type="RollingUpdate", status.currentNumberScheduled="0.000000", status.desiredNumberScheduled="0.000000", status.numberMisscheduled="0.000000", status.numberReady="0.000000" 2025-12-04T12:01:39.911577047Z I1204 12:01:39.909851 1 controller.go:340] object changed: *v1.Config, Name=cluster (status=true): changed:status.conditions.3.lastTransitionTime={"2025-12-04T12:00:23Z" -> "2025-12-04T12:01:39Z"} 2025-12-04T12:01:39.914934573Z I1204 12:01:39.912373 1 generator.go:63] object *v1.ClusterOperator, Name=image-registry updated: changed:metadata.managedFields.2.time={"2025-12-04T12:00:04Z" -> "2025-12-04T12:01:39Z"}, changed:metadata.resourceVersion={"12192" -> "14288"}, changed:status.conditions.0.lastTransitionTime={"2025-12-04T12:00:01Z" -> "2025-12-04T12:01:39Z"}, changed:status.conditions.0.message={"Available: The registry is removed\nImagePrunerAvailable: Pruner CronJob has been created" -> "Available: The registry is removed\nNodeCADaemonAvailable: The daemon set node-ca does not exist\nImagePrunerAvailable: Pruner CronJob has been created"}, changed:status.conditions.0.reason={"Removed" -> "NodeCADaemonNotFound::Removed"}, changed:status.conditions.0.status={"True" -> "False"}, changed:status.conditions.1.lastTransitionTime={"2025-12-04T12:00:01Z" -> "2025-12-04T12:01:39Z"}, changed:status.conditions.1.message={"Progressing: All registry resources are removed" -> "Progressing: All registry resources are removed\nNodeCADaemonProgressing: The daemon set node-ca does not exist"}, changed:status.conditions.1.reason={"Removed" -> "NodeCADaemonNotFound::Removed"}, changed:status.conditions.1.status={"False" -> "True"} 2025-12-04T12:01:39.947748091Z E1204 12:01:39.947052 1 controller.go:379] unable to sync: Operation cannot be fulfilled on configs.imageregistry.operator.openshift.io "cluster": the object has been modified; please apply your changes to the latest version and try again, requeuing 2025-12-04T12:01:39.957499300Z I1204 12:01:39.957350 1 generator.go:63] object *v1.ClusterOperator, Name=image-registry updated: changed:metadata.resourceVersion={"14288" -> "14290"}, changed:status.conditions.0.message={"Available: The registry is removed\nNodeCADaemonAvailable: The daemon set node-ca does not exist\nImagePrunerAvailable: Pruner CronJob has been created" -> "Available: The registry is removed\nNodeCADaemonAvailable: The daemon set node-ca does not have available replicas\nImagePrunerAvailable: Pruner CronJob has been created"}, changed:status.conditions.0.reason={"NodeCADaemonNotFound::Removed" -> "NodeCADaemonNoAvailableReplicas::Removed"}, changed:status.conditions.1.message={"Progressing: All registry resources are removed\nNodeCADaemonProgressing: The daemon set node-ca does not exist" -> "Progressing: All registry resources are removed\nNodeCADaemonProgressing: The daemon set node-ca is updating node pods"}, changed:status.conditions.1.reason={"NodeCADaemonNotFound::Removed" -> "NodeCADaemonProgressing::Removed"} 2025-12-04T12:01:39.988009576Z I1204 12:01:39.987556 1 controller.go:340] object changed: *v1.Config, Name=cluster (status=true): changed:status.conditions.3.lastTransitionTime={"2025-12-04T12:00:23Z" -> "2025-12-04T12:01:39Z"} 2025-12-04T12:01:40.417350932Z I1204 12:01:40.417249 1 controller.go:340] object changed: *v1.Config, Name=cluster (status=true): changed:status.conditions.3.lastTransitionTime={"2025-12-04T12:01:39Z" -> "2025-12-04T12:01:40Z"} 2025-12-04T12:01:41.215076696Z I1204 12:01:41.214991 1 controller.go:340] object changed: *v1.Config, Name=cluster (status=true): changed:status.conditions.3.lastTransitionTime={"2025-12-04T12:01:40Z" -> "2025-12-04T12:01:41Z"} 2025-12-04T12:01:42.019389868Z I1204 12:01:42.019286 1 controller.go:340] object changed: *v1.Config, Name=cluster (status=true): changed:status.conditions.3.lastTransitionTime={"2025-12-04T12:01:41Z" -> "2025-12-04T12:01:42Z"} 2025-12-04T12:01:42.819542109Z I1204 12:01:42.819444 1 controller.go:340] object changed: *v1.Config, Name=cluster (status=true): cluster-image-registry-operator_previous.log0000640000000000000000000006166015114274105040714 0ustar0000000000000000config/pod/openshift-image-registry/logs/cluster-image-registry-operator-6fb9f88b7-tgvfl2025-12-04T11:50:35.770025450Z W1204 11:50:35.769940 1 reflector.go:561] github.com/openshift/client-go/route/informers/externalversions/factory.go:125: failed to list *v1.Route: the server is currently unable to handle the request (get routes.route.openshift.io) 2025-12-04T11:50:35.770025450Z E1204 11:50:35.770005 1 reflector.go:158] "Unhandled Error" err="github.com/openshift/client-go/route/informers/externalversions/factory.go:125: Failed to watch *v1.Route: failed to list *v1.Route: the server is currently unable to handle the request (get routes.route.openshift.io)" logger="UnhandledError" 2025-12-04T11:50:36.229283108Z I1204 11:50:36.229181 1 caconfig.go:127] unable to get the service name to add service-ca.crt 2025-12-04T11:50:36.229283108Z E1204 11:50:36.229216 1 clusteroperator.go:104] unable to sync ClusterOperatorStatusController: config.imageregistry.operator.openshift.io "cluster" not found, requeuing 2025-12-04T11:50:36.229366100Z I1204 11:50:36.229297 1 imageregistryca.go:130] unable to get the service name to add service-ca.crt 2025-12-04T11:50:36.229366100Z E1204 11:50:36.229336 1 imageregistrycertificates.go:144] ImageRegistryCertificatesController: unable to sync: config.imageregistry.operator.openshift.io "cluster" not found, requeuing 2025-12-04T11:50:36.229366100Z E1204 11:50:36.229294 1 azurestackcloud.go:78] AzureStackCloudController: unable to sync: config.imageregistry.operator.openshift.io "cluster" not found, requeuing 2025-12-04T11:50:36.235315150Z E1204 11:50:36.235229 1 nodecadaemon.go:105] NodeCADaemonController: unable to sync: [failed to create object *v1.DaemonSet, Namespace=openshift-image-registry, Name=node-ca: config.imageregistry.operator.openshift.io "cluster" not found, config.imageregistry.operator.openshift.io "cluster" not found], requeuing 2025-12-04T11:50:36.371947322Z W1204 11:50:36.371828 1 reflector.go:561] github.com/openshift/client-go/image/informers/externalversions/factory.go:125: failed to list *v1.ImageStream: the server is currently unable to handle the request (get imagestreams.image.openshift.io) 2025-12-04T11:50:36.372008284Z E1204 11:50:36.371949 1 reflector.go:158] "Unhandled Error" err="github.com/openshift/client-go/image/informers/externalversions/factory.go:125: Failed to watch *v1.ImageStream: failed to list *v1.ImageStream: the server is currently unable to handle the request (get imagestreams.image.openshift.io)" logger="UnhandledError" 2025-12-04T11:50:36.386969016Z E1204 11:50:36.383998 1 clusteroperator.go:104] unable to sync ClusterOperatorStatusController: config.imageregistry.operator.openshift.io "cluster" not found, requeuing 2025-12-04T11:50:37.509747745Z I1204 11:50:37.509641 1 caconfig.go:127] unable to get the service name to add service-ca.crt 2025-12-04T11:50:37.509747745Z I1204 11:50:37.509727 1 imageregistryca.go:130] unable to get the service name to add service-ca.crt 2025-12-04T11:50:37.509862098Z E1204 11:50:37.509756 1 imageregistrycertificates.go:144] ImageRegistryCertificatesController: unable to sync: config.imageregistry.operator.openshift.io "cluster" not found, requeuing 2025-12-04T11:50:37.509924730Z E1204 11:50:37.509886 1 azurestackcloud.go:78] AzureStackCloudController: unable to sync: config.imageregistry.operator.openshift.io "cluster" not found, requeuing 2025-12-04T11:50:37.516870849Z E1204 11:50:37.516782 1 nodecadaemon.go:105] NodeCADaemonController: unable to sync: [failed to create object *v1.DaemonSet, Namespace=openshift-image-registry, Name=node-ca: config.imageregistry.operator.openshift.io "cluster" not found, config.imageregistry.operator.openshift.io "cluster" not found], requeuing 2025-12-04T11:50:37.575429586Z W1204 11:50:37.575306 1 reflector.go:561] github.com/openshift/client-go/route/informers/externalversions/factory.go:125: failed to list *v1.Route: the server is currently unable to handle the request (get routes.route.openshift.io) 2025-12-04T11:50:37.575429586Z E1204 11:50:37.575378 1 reflector.go:158] "Unhandled Error" err="github.com/openshift/client-go/route/informers/externalversions/factory.go:125: Failed to watch *v1.Route: failed to list *v1.Route: the server is currently unable to handle the request (get routes.route.openshift.io)" logger="UnhandledError" 2025-12-04T11:50:38.328881591Z W1204 11:50:38.328774 1 reflector.go:561] github.com/openshift/client-go/image/informers/externalversions/factory.go:125: failed to list *v1.ImageStream: the server is currently unable to handle the request (get imagestreams.image.openshift.io) 2025-12-04T11:50:38.328944513Z E1204 11:50:38.328889 1 reflector.go:158] "Unhandled Error" err="github.com/openshift/client-go/image/informers/externalversions/factory.go:125: Failed to watch *v1.ImageStream: failed to list *v1.ImageStream: the server is currently unable to handle the request (get imagestreams.image.openshift.io)" logger="UnhandledError" 2025-12-04T11:50:38.790346565Z E1204 11:50:38.790236 1 clusteroperator.go:104] unable to sync ClusterOperatorStatusController: config.imageregistry.operator.openshift.io "cluster" not found, requeuing 2025-12-04T11:50:39.251766718Z E1204 11:50:39.250976 1 clusteroperator.go:104] unable to sync ClusterOperatorStatusController: config.imageregistry.operator.openshift.io "cluster" not found, requeuing 2025-12-04T11:50:39.483307285Z E1204 11:50:39.483231 1 clusteroperator.go:104] unable to sync ClusterOperatorStatusController: config.imageregistry.operator.openshift.io "cluster" not found, requeuing 2025-12-04T11:50:39.637075315Z E1204 11:50:39.636961 1 clusteroperator.go:104] unable to sync ClusterOperatorStatusController: config.imageregistry.operator.openshift.io "cluster" not found, requeuing 2025-12-04T11:50:39.969924778Z E1204 11:50:39.966137 1 clusteroperator.go:104] unable to sync ClusterOperatorStatusController: config.imageregistry.operator.openshift.io "cluster" not found, requeuing 2025-12-04T11:50:39.994374765Z E1204 11:50:39.994286 1 clusteroperator.go:104] unable to sync ClusterOperatorStatusController: config.imageregistry.operator.openshift.io "cluster" not found, requeuing 2025-12-04T11:50:40.041907170Z E1204 11:50:40.041785 1 clusteroperator.go:104] unable to sync ClusterOperatorStatusController: config.imageregistry.operator.openshift.io "cluster" not found, requeuing 2025-12-04T11:50:40.071004518Z I1204 11:50:40.070908 1 caconfig.go:127] unable to get the service name to add service-ca.crt 2025-12-04T11:50:40.071044199Z I1204 11:50:40.071009 1 imageregistryca.go:130] unable to get the service name to add service-ca.crt 2025-12-04T11:50:40.071044199Z E1204 11:50:40.071009 1 azurestackcloud.go:78] AzureStackCloudController: unable to sync: config.imageregistry.operator.openshift.io "cluster" not found, requeuing 2025-12-04T11:50:40.071070040Z E1204 11:50:40.071044 1 imageregistrycertificates.go:144] ImageRegistryCertificatesController: unable to sync: config.imageregistry.operator.openshift.io "cluster" not found, requeuing 2025-12-04T11:50:40.077861225Z E1204 11:50:40.077762 1 nodecadaemon.go:105] NodeCADaemonController: unable to sync: [failed to create object *v1.DaemonSet, Namespace=openshift-image-registry, Name=node-ca: config.imageregistry.operator.openshift.io "cluster" not found, config.imageregistry.operator.openshift.io "cluster" not found], requeuing 2025-12-04T11:50:40.130157523Z E1204 11:50:40.130046 1 clusteroperator.go:104] unable to sync ClusterOperatorStatusController: config.imageregistry.operator.openshift.io "cluster" not found, requeuing 2025-12-04T11:50:40.816747550Z E1204 11:50:40.816351 1 clusteroperator.go:104] unable to sync ClusterOperatorStatusController: config.imageregistry.operator.openshift.io "cluster" not found, requeuing 2025-12-04T11:50:41.288665330Z W1204 11:50:41.288570 1 reflector.go:561] github.com/openshift/client-go/route/informers/externalversions/factory.go:125: failed to list *v1.Route: the server is currently unable to handle the request (get routes.route.openshift.io) 2025-12-04T11:50:41.288712482Z E1204 11:50:41.288658 1 reflector.go:158] "Unhandled Error" err="github.com/openshift/client-go/route/informers/externalversions/factory.go:125: Failed to watch *v1.Route: failed to list *v1.Route: the server is currently unable to handle the request (get routes.route.openshift.io)" logger="UnhandledError" 2025-12-04T11:50:42.910411295Z W1204 11:50:42.910325 1 reflector.go:561] github.com/openshift/client-go/image/informers/externalversions/factory.go:125: failed to list *v1.ImageStream: the server is currently unable to handle the request (get imagestreams.image.openshift.io) 2025-12-04T11:50:42.910411295Z E1204 11:50:42.910378 1 reflector.go:158] "Unhandled Error" err="github.com/openshift/client-go/image/informers/externalversions/factory.go:125: Failed to watch *v1.ImageStream: failed to list *v1.ImageStream: the server is currently unable to handle the request (get imagestreams.image.openshift.io)" logger="UnhandledError" 2025-12-04T11:50:45.191225976Z I1204 11:50:45.191123 1 caconfig.go:127] unable to get the service name to add service-ca.crt 2025-12-04T11:50:45.191225976Z E1204 11:50:45.191157 1 azurestackcloud.go:78] AzureStackCloudController: unable to sync: config.imageregistry.operator.openshift.io "cluster" not found, requeuing 2025-12-04T11:50:45.191225976Z I1204 11:50:45.191210 1 imageregistryca.go:130] unable to get the service name to add service-ca.crt 2025-12-04T11:50:45.191308489Z E1204 11:50:45.191232 1 imageregistrycertificates.go:144] ImageRegistryCertificatesController: unable to sync: config.imageregistry.operator.openshift.io "cluster" not found, requeuing 2025-12-04T11:50:45.199037862Z E1204 11:50:45.198968 1 nodecadaemon.go:105] NodeCADaemonController: unable to sync: [failed to create object *v1.DaemonSet, Namespace=openshift-image-registry, Name=node-ca: config.imageregistry.operator.openshift.io "cluster" not found, config.imageregistry.operator.openshift.io "cluster" not found], requeuing 2025-12-04T11:50:48.011527826Z W1204 11:50:48.011426 1 reflector.go:561] github.com/openshift/client-go/route/informers/externalversions/factory.go:125: failed to list *v1.Route: the server is currently unable to handle the request (get routes.route.openshift.io) 2025-12-04T11:50:48.011527826Z E1204 11:50:48.011503 1 reflector.go:158] "Unhandled Error" err="github.com/openshift/client-go/route/informers/externalversions/factory.go:125: Failed to watch *v1.Route: failed to list *v1.Route: the server is currently unable to handle the request (get routes.route.openshift.io)" logger="UnhandledError" 2025-12-04T11:50:49.030673238Z E1204 11:50:49.030582 1 clusteroperator.go:104] unable to sync ClusterOperatorStatusController: config.imageregistry.operator.openshift.io "cluster" not found, requeuing 2025-12-04T11:50:53.239334661Z W1204 11:50:53.239245 1 reflector.go:561] github.com/openshift/client-go/image/informers/externalversions/factory.go:125: failed to list *v1.ImageStream: the server is currently unable to handle the request (get imagestreams.image.openshift.io) 2025-12-04T11:50:53.239334661Z E1204 11:50:53.239311 1 reflector.go:158] "Unhandled Error" err="github.com/openshift/client-go/image/informers/externalversions/factory.go:125: Failed to watch *v1.ImageStream: failed to list *v1.ImageStream: the server is currently unable to handle the request (get imagestreams.image.openshift.io)" logger="UnhandledError" 2025-12-04T11:50:55.432008192Z I1204 11:50:55.431896 1 caconfig.go:127] unable to get the service name to add service-ca.crt 2025-12-04T11:50:55.432008192Z I1204 11:50:55.431997 1 imageregistryca.go:130] unable to get the service name to add service-ca.crt 2025-12-04T11:50:55.432092585Z E1204 11:50:55.432015 1 azurestackcloud.go:78] AzureStackCloudController: unable to sync: config.imageregistry.operator.openshift.io "cluster" not found, requeuing 2025-12-04T11:50:55.432092585Z E1204 11:50:55.432037 1 imageregistrycertificates.go:144] ImageRegistryCertificatesController: unable to sync: config.imageregistry.operator.openshift.io "cluster" not found, requeuing 2025-12-04T11:50:55.439991083Z E1204 11:50:55.439903 1 nodecadaemon.go:105] NodeCADaemonController: unable to sync: [failed to create object *v1.DaemonSet, Namespace=openshift-image-registry, Name=node-ca: config.imageregistry.operator.openshift.io "cluster" not found, config.imageregistry.operator.openshift.io "cluster" not found], requeuing 2025-12-04T11:51:10.197615031Z W1204 11:51:10.197527 1 reflector.go:561] github.com/openshift/client-go/route/informers/externalversions/factory.go:125: failed to list *v1.Route: the server is currently unable to handle the request (get routes.route.openshift.io) 2025-12-04T11:51:10.197615031Z E1204 11:51:10.197602 1 reflector.go:158] "Unhandled Error" err="github.com/openshift/client-go/route/informers/externalversions/factory.go:125: Failed to watch *v1.Route: failed to list *v1.Route: the server is currently unable to handle the request (get routes.route.openshift.io)" logger="UnhandledError" 2025-12-04T11:51:15.912736559Z I1204 11:51:15.912637 1 caconfig.go:127] unable to get the service name to add service-ca.crt 2025-12-04T11:51:15.912736559Z E1204 11:51:15.912699 1 azurestackcloud.go:78] AzureStackCloudController: unable to sync: config.imageregistry.operator.openshift.io "cluster" not found, requeuing 2025-12-04T11:51:15.912829262Z I1204 11:51:15.912735 1 imageregistryca.go:130] unable to get the service name to add service-ca.crt 2025-12-04T11:51:15.912829262Z E1204 11:51:15.912756 1 imageregistrycertificates.go:144] ImageRegistryCertificatesController: unable to sync: config.imageregistry.operator.openshift.io "cluster" not found, requeuing 2025-12-04T11:51:15.921693140Z E1204 11:51:15.921581 1 nodecadaemon.go:105] NodeCADaemonController: unable to sync: [failed to create object *v1.DaemonSet, Namespace=openshift-image-registry, Name=node-ca: config.imageregistry.operator.openshift.io "cluster" not found, config.imageregistry.operator.openshift.io "cluster" not found], requeuing 2025-12-04T11:51:16.325984300Z W1204 11:51:16.325906 1 reflector.go:561] github.com/openshift/client-go/image/informers/externalversions/factory.go:125: failed to list *v1.ImageStream: the server is currently unable to handle the request (get imagestreams.image.openshift.io) 2025-12-04T11:51:16.326023841Z E1204 11:51:16.325973 1 reflector.go:158] "Unhandled Error" err="github.com/openshift/client-go/image/informers/externalversions/factory.go:125: Failed to watch *v1.ImageStream: failed to list *v1.ImageStream: the server is currently unable to handle the request (get imagestreams.image.openshift.io)" logger="UnhandledError" 2025-12-04T11:51:40.320188551Z W1204 11:51:40.320020 1 reflector.go:561] github.com/openshift/client-go/route/informers/externalversions/factory.go:125: failed to list *v1.Route: the server is currently unable to handle the request (get routes.route.openshift.io) 2025-12-04T11:51:40.320188551Z E1204 11:51:40.320128 1 reflector.go:158] "Unhandled Error" err="github.com/openshift/client-go/route/informers/externalversions/factory.go:125: Failed to watch *v1.Route: failed to list *v1.Route: the server is currently unable to handle the request (get routes.route.openshift.io)" logger="UnhandledError" 2025-12-04T11:51:55.886775220Z W1204 11:51:55.886660 1 reflector.go:561] github.com/openshift/client-go/image/informers/externalversions/factory.go:125: failed to list *v1.ImageStream: the server is currently unable to handle the request (get imagestreams.image.openshift.io) 2025-12-04T11:51:55.886775220Z E1204 11:51:55.886757 1 reflector.go:158] "Unhandled Error" err="github.com/openshift/client-go/image/informers/externalversions/factory.go:125: Failed to watch *v1.ImageStream: failed to list *v1.ImageStream: the server is currently unable to handle the request (get imagestreams.image.openshift.io)" logger="UnhandledError" 2025-12-04T11:51:56.873925529Z I1204 11:51:56.873607 1 caconfig.go:127] unable to get the service name to add service-ca.crt 2025-12-04T11:51:56.873925529Z E1204 11:51:56.873662 1 azurestackcloud.go:78] AzureStackCloudController: unable to sync: config.imageregistry.operator.openshift.io "cluster" not found, requeuing 2025-12-04T11:51:56.873925529Z I1204 11:51:56.873726 1 imageregistryca.go:130] unable to get the service name to add service-ca.crt 2025-12-04T11:51:56.873925529Z E1204 11:51:56.873758 1 imageregistrycertificates.go:144] ImageRegistryCertificatesController: unable to sync: config.imageregistry.operator.openshift.io "cluster" not found, requeuing 2025-12-04T11:51:56.882690411Z E1204 11:51:56.882571 1 nodecadaemon.go:105] NodeCADaemonController: unable to sync: [failed to create object *v1.DaemonSet, Namespace=openshift-image-registry, Name=node-ca: config.imageregistry.operator.openshift.io "cluster" not found, config.imageregistry.operator.openshift.io "cluster" not found], requeuing 2025-12-04T11:52:08.845040812Z E1204 11:52:08.844884 1 leaderelection.go:429] Failed to update lock optimitically: Timeout: request did not complete within requested timeout - context deadline exceeded, falling back to slow path 2025-12-04T11:52:31.360344123Z W1204 11:52:31.360273 1 reflector.go:561] github.com/openshift/client-go/image/informers/externalversions/factory.go:125: failed to list *v1.ImageStream: the server is currently unable to handle the request (get imagestreams.image.openshift.io) 2025-12-04T11:52:31.360344123Z E1204 11:52:31.360327 1 reflector.go:158] "Unhandled Error" err="github.com/openshift/client-go/image/informers/externalversions/factory.go:125: Failed to watch *v1.ImageStream: failed to list *v1.ImageStream: the server is currently unable to handle the request (get imagestreams.image.openshift.io)" logger="UnhandledError" 2025-12-04T11:52:34.690229590Z W1204 11:52:34.690099 1 reflector.go:561] github.com/openshift/client-go/route/informers/externalversions/factory.go:125: failed to list *v1.Route: the server is currently unable to handle the request (get routes.route.openshift.io) 2025-12-04T11:52:34.690229590Z E1204 11:52:34.690171 1 reflector.go:158] "Unhandled Error" err="github.com/openshift/client-go/route/informers/externalversions/factory.go:125: Failed to watch *v1.Route: failed to list *v1.Route: the server is currently unable to handle the request (get routes.route.openshift.io)" logger="UnhandledError" 2025-12-04T11:53:08.846596929Z E1204 11:53:08.846507 1 leaderelection.go:436] error retrieving resource lock openshift-image-registry/openshift-master-controllers: the server was unable to return a response in the time allotted, but may still be processing the request (get leases.coordination.k8s.io openshift-master-controllers) 2025-12-04T11:53:18.794694999Z I1204 11:53:18.794596 1 caconfig.go:127] unable to get the service name to add service-ca.crt 2025-12-04T11:53:18.795309229Z I1204 11:53:18.794724 1 imageregistryca.go:130] unable to get the service name to add service-ca.crt 2025-12-04T11:53:18.795309229Z E1204 11:53:18.794752 1 imageregistrycertificates.go:144] ImageRegistryCertificatesController: unable to sync: config.imageregistry.operator.openshift.io "cluster" not found, requeuing 2025-12-04T11:53:18.795309229Z E1204 11:53:18.794863 1 azurestackcloud.go:78] AzureStackCloudController: unable to sync: config.imageregistry.operator.openshift.io "cluster" not found, requeuing 2025-12-04T11:53:18.803593214Z E1204 11:53:18.803520 1 nodecadaemon.go:105] NodeCADaemonController: unable to sync: [failed to create object *v1.DaemonSet, Namespace=openshift-image-registry, Name=node-ca: config.imageregistry.operator.openshift.io "cluster" not found, config.imageregistry.operator.openshift.io "cluster" not found], requeuing 2025-12-04T11:53:23.675225460Z W1204 11:53:23.675137 1 reflector.go:561] github.com/openshift/client-go/image/informers/externalversions/factory.go:125: failed to list *v1.ImageStream: the server is currently unable to handle the request (get imagestreams.image.openshift.io) 2025-12-04T11:53:23.675225460Z E1204 11:53:23.675202 1 reflector.go:158] "Unhandled Error" err="github.com/openshift/client-go/image/informers/externalversions/factory.go:125: Failed to watch *v1.ImageStream: failed to list *v1.ImageStream: the server is currently unable to handle the request (get imagestreams.image.openshift.io)" logger="UnhandledError" 2025-12-04T11:53:32.139354494Z W1204 11:53:32.139269 1 reflector.go:561] github.com/openshift/client-go/route/informers/externalversions/factory.go:125: failed to list *v1.Route: the server is currently unable to handle the request (get routes.route.openshift.io) 2025-12-04T11:53:32.139354494Z E1204 11:53:32.139332 1 reflector.go:158] "Unhandled Error" err="github.com/openshift/client-go/route/informers/externalversions/factory.go:125: Failed to watch *v1.Route: failed to list *v1.Route: the server is currently unable to handle the request (get routes.route.openshift.io)" logger="UnhandledError" 2025-12-04T11:54:10.690140240Z W1204 11:54:10.690046 1 reflector.go:561] github.com/openshift/client-go/image/informers/externalversions/factory.go:125: failed to list *v1.ImageStream: the server is currently unable to handle the request (get imagestreams.image.openshift.io) 2025-12-04T11:54:10.690140240Z E1204 11:54:10.690120 1 reflector.go:158] "Unhandled Error" err="github.com/openshift/client-go/image/informers/externalversions/factory.go:125: Failed to watch *v1.ImageStream: failed to list *v1.ImageStream: the server is currently unable to handle the request (get imagestreams.image.openshift.io)" logger="UnhandledError" 2025-12-04T11:54:13.839877307Z W1204 11:54:13.839750 1 reflector.go:561] github.com/openshift/client-go/route/informers/externalversions/factory.go:125: failed to list *v1.Route: the server is currently unable to handle the request (get routes.route.openshift.io) 2025-12-04T11:54:13.839877307Z E1204 11:54:13.839836 1 reflector.go:158] "Unhandled Error" err="github.com/openshift/client-go/route/informers/externalversions/factory.go:125: Failed to watch *v1.Route: failed to list *v1.Route: the server is currently unable to handle the request (get routes.route.openshift.io)" logger="UnhandledError" 2025-12-04T11:54:42.851033485Z E1204 11:54:42.850629 1 leaderelection.go:429] Failed to update lock optimitically: Timeout: request did not complete within requested timeout - context deadline exceeded, falling back to slow path 2025-12-04T11:54:49.472617092Z W1204 11:54:49.472512 1 reflector.go:561] github.com/openshift/client-go/route/informers/externalversions/factory.go:125: failed to list *v1.Route: the server is currently unable to handle the request (get routes.route.openshift.io) 2025-12-04T11:54:49.472617092Z E1204 11:54:49.472587 1 reflector.go:158] "Unhandled Error" err="github.com/openshift/client-go/route/informers/externalversions/factory.go:125: Failed to watch *v1.Route: failed to list *v1.Route: the server is currently unable to handle the request (get routes.route.openshift.io)" logger="UnhandledError" 2025-12-04T11:55:00.509323972Z W1204 11:55:00.509250 1 reflector.go:561] github.com/openshift/client-go/image/informers/externalversions/factory.go:125: failed to list *v1.ImageStream: the server is currently unable to handle the request (get imagestreams.image.openshift.io) 2025-12-04T11:55:00.509477417Z E1204 11:55:00.509457 1 reflector.go:158] "Unhandled Error" err="github.com/openshift/client-go/image/informers/externalversions/factory.go:125: Failed to watch *v1.ImageStream: failed to list *v1.ImageStream: the server is currently unable to handle the request (get imagestreams.image.openshift.io)" logger="UnhandledError" 2025-12-04T11:55:34.842258361Z E1204 11:55:34.842166 1 leaderelection.go:436] error retrieving resource lock openshift-image-registry/openshift-master-controllers: Get "https://172.30.0.1:443/apis/coordination.k8s.io/v1/namespaces/openshift-image-registry/leases/openshift-master-controllers?timeout=4m0s": context deadline exceeded 2025-12-04T11:55:34.842258361Z I1204 11:55:34.842232 1 leaderelection.go:297] failed to renew lease openshift-image-registry/openshift-master-controllers: timed out waiting for the condition 2025-12-04T11:55:46.356792605Z W1204 11:55:46.356717 1 reflector.go:561] github.com/openshift/client-go/image/informers/externalversions/factory.go:125: failed to list *v1.ImageStream: the server is currently unable to handle the request (get imagestreams.image.openshift.io) 2025-12-04T11:55:46.356792605Z E1204 11:55:46.356770 1 reflector.go:158] "Unhandled Error" err="github.com/openshift/client-go/image/informers/externalversions/factory.go:125: Failed to watch *v1.ImageStream: failed to list *v1.ImageStream: the server is currently unable to handle the request (get imagestreams.image.openshift.io)" logger="UnhandledError" 2025-12-04T11:55:46.782669588Z W1204 11:55:46.782579 1 leaderelection.go:84] leader election lost config/pod/openshift-image-registry/cluster-image-registry-operator-6fb9f88b7-tgvfl.json0000640000000000000000000001602515114274105030016 0ustar0000000000000000{"metadata":{"name":"cluster-image-registry-operator-6fb9f88b7-tgvfl","generateName":"cluster-image-registry-operator-6fb9f88b7-","namespace":"openshift-image-registry","uid":"1ade9186-b262-4b5e-9c6f-ba89f0130fb6","resourceVersion":"11265","creationTimestamp":"2025-12-04T11:35:49Z","labels":{"name":"cluster-image-registry-operator","pod-template-hash":"6fb9f88b7"},"annotations":{"k8s.ovn.org/pod-networks":"{\"default\":{\"ip_addresses\":[\"10.128.0.24/23\"],\"mac_address\":\"0a:58:0a:80:00:18\",\"gateway_ips\":[\"10.128.0.1\"],\"routes\":[{\"dest\":\"10.128.0.0/16\",\"nextHop\":\"10.128.0.1\"},{\"dest\":\"172.30.0.0/16\",\"nextHop\":\"10.128.0.1\"},{\"dest\":\"169.254.0.5/32\",\"nextHop\":\"10.128.0.1\"},{\"dest\":\"100.64.0.0/16\",\"nextHop\":\"10.128.0.1\"}],\"ip_address\":\"10.128.0.24/23\",\"gateway_ip\":\"10.128.0.1\",\"role\":\"primary\"}}","k8s.v1.cni.cncf.io/network-status":"[{\n \"name\": \"ovn-kubernetes\",\n \"interface\": \"eth0\",\n \"ips\": [\n \"10.128.0.24\"\n ],\n \"mac\": \"0a:58:0a:80:00:18\",\n \"default\": true,\n \"dns\": {}\n}]","openshift.io/required-scc":"restricted-v2","openshift.io/scc":"restricted-v2","seccomp.security.alpha.kubernetes.io/pod":"runtime/default"},"ownerReferences":[{"apiVersion":"apps/v1","kind":"ReplicaSet","name":"cluster-image-registry-operator-6fb9f88b7","uid":"6d4aa020-6861-419f-8069-63186e428118","controller":true,"blockOwnerDeletion":true}]},"spec":{"volumes":[{"name":"image-registry-operator-tls","secret":{"secretName":"image-registry-operator-tls","defaultMode":420}},{"name":"trusted-ca","configMap":{"name":"trusted-ca","items":[{"key":"ca-bundle.crt","path":"tls-ca-bundle.pem"}],"defaultMode":420,"optional":true}},{"name":"bound-sa-token","projected":{"sources":[{"serviceAccountToken":{"audience":"openshift","expirationSeconds":3600,"path":"token"}}],"defaultMode":420}},{"name":"kube-api-access-t46k7","projected":{"sources":[{"serviceAccountToken":{"expirationSeconds":3607,"path":"token"}},{"configMap":{"name":"kube-root-ca.crt","items":[{"key":"ca.crt","path":"ca.crt"}]}},{"downwardAPI":{"items":[{"path":"namespace","fieldRef":{"apiVersion":"v1","fieldPath":"metadata.namespace"}}]}},{"configMap":{"name":"openshift-service-ca.crt","items":[{"key":"service-ca.crt","path":"service-ca.crt"}]}}],"defaultMode":420}}],"containers":[{"name":"cluster-image-registry-operator","image":"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:aa24edce3d740f84c40018e94cdbf2bc7375268d13d57c2d664e43a46ccea3fc","args":["--files=/var/run/configmaps/trusted-ca/tls-ca-bundle.pem","--files=/etc/secrets/tls.crt","--files=/etc/secrets/tls.key"],"ports":[{"name":"metrics","containerPort":60000,"protocol":"TCP"}],"env":[{"name":"RELEASE_VERSION","value":"4.18.29"},{"name":"WATCH_NAMESPACE","valueFrom":{"fieldRef":{"apiVersion":"v1","fieldPath":"metadata.namespace"}}},{"name":"POD_NAME","valueFrom":{"fieldRef":{"apiVersion":"v1","fieldPath":"metadata.name"}}},{"name":"OPERATOR_NAME","value":"cluster-image-registry-operator"},{"name":"OPERATOR_IMAGE","value":"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:aa24edce3d740f84c40018e94cdbf2bc7375268d13d57c2d664e43a46ccea3fc"},{"name":"IMAGE","value":"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3ebe19b23694155a15d0968968fdee3dcf200ab9718ae1fcbd05f4d24960b827"},{"name":"IMAGE_PRUNER","value":"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:79f99fd6cce984287932edf0d009660bb488d663081f3d62ec3b23bc8bfbf6c2"},{"name":"AZURE_ENVIRONMENT_FILEPATH","value":"/tmp/azurestackcloud.json"},{"name":"OPERATOR_IMAGE_VERSION","value":"4.18.29"}],"resources":{"requests":{"cpu":"10m","memory":"50Mi"}},"volumeMounts":[{"name":"trusted-ca","mountPath":"/var/run/configmaps/trusted-ca/"},{"name":"image-registry-operator-tls","mountPath":"/etc/secrets"},{"name":"bound-sa-token","readOnly":true,"mountPath":"/var/run/secrets/openshift/serviceaccount"},{"name":"kube-api-access-t46k7","readOnly":true,"mountPath":"/var/run/secrets/kubernetes.io/serviceaccount"}],"terminationMessagePath":"/dev/termination-log","terminationMessagePolicy":"FallbackToLogsOnError","imagePullPolicy":"IfNotPresent","securityContext":{"capabilities":{"drop":["ALL"]},"runAsUser":1000300000,"runAsNonRoot":true,"allowPrivilegeEscalation":false}}],"restartPolicy":"Always","terminationGracePeriodSeconds":30,"dnsPolicy":"ClusterFirst","nodeSelector":{"node-role.kubernetes.io/master":""},"serviceAccountName":"cluster-image-registry-operator","serviceAccount":"cluster-image-registry-operator","nodeName":"master-0","shareProcessNamespace":false,"securityContext":{"seLinuxOptions":{"level":"s0:c17,c14"},"fsGroup":1000300000,"seccompProfile":{"type":"RuntimeDefault"}},"schedulerName":"default-scheduler","tolerations":[{"key":"node-role.kubernetes.io/master","operator":"Exists","effect":"NoSchedule"},{"key":"node.kubernetes.io/unreachable","operator":"Exists","effect":"NoExecute","tolerationSeconds":120},{"key":"node.kubernetes.io/not-ready","operator":"Exists","effect":"NoExecute","tolerationSeconds":120},{"key":"node.kubernetes.io/memory-pressure","operator":"Exists","effect":"NoSchedule"}],"priorityClassName":"system-cluster-critical","priority":2000000000,"enableServiceLinks":true,"preemptionPolicy":"PreemptLowerPriority"},"status":{"phase":"Running","conditions":[{"type":"PodReadyToStartContainers","status":"True","lastProbeTime":null,"lastTransitionTime":"2025-12-04T11:38:19Z"},{"type":"Initialized","status":"True","lastProbeTime":null,"lastTransitionTime":"2025-12-04T11:37:30Z"},{"type":"Ready","status":"True","lastProbeTime":null,"lastTransitionTime":"2025-12-04T11:55:48Z"},{"type":"ContainersReady","status":"True","lastProbeTime":null,"lastTransitionTime":"2025-12-04T11:55:48Z"},{"type":"PodScheduled","status":"True","lastProbeTime":null,"lastTransitionTime":"2025-12-04T11:37:30Z"}],"hostIP":"192.168.32.10","hostIPs":[{"ip":"192.168.32.10"}],"podIP":"10.128.0.24","podIPs":[{"ip":"10.128.0.24"}],"startTime":"2025-12-04T11:37:30Z","containerStatuses":[{"name":"cluster-image-registry-operator","state":{"running":{"startedAt":"2025-12-04T11:55:47Z"}},"lastState":{"terminated":{"exitCode":0,"reason":"Completed","startedAt":"2025-12-04T11:45:02Z","finishedAt":"2025-12-04T11:55:46Z","containerID":"cri-o://c1818ca43db2765bb639a6e97b3c2f8506a31edb0b1a76aecba6684bf041bcbf"}},"ready":true,"restartCount":2,"image":"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:aa24edce3d740f84c40018e94cdbf2bc7375268d13d57c2d664e43a46ccea3fc","imageID":"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:aa24edce3d740f84c40018e94cdbf2bc7375268d13d57c2d664e43a46ccea3fc","containerID":"cri-o://ccf27b7abfb6528c74fd1f7c21da05e349a9bda10d36d097585e26643cfe95b2","started":true,"volumeMounts":[{"name":"trusted-ca","mountPath":"/var/run/configmaps/trusted-ca/"},{"name":"image-registry-operator-tls","mountPath":"/etc/secrets"},{"name":"bound-sa-token","mountPath":"/var/run/secrets/openshift/serviceaccount","readOnly":true,"recursiveReadOnly":"Disabled"},{"name":"kube-api-access-t46k7","mountPath":"/var/run/secrets/kubernetes.io/serviceaccount","readOnly":true,"recursiveReadOnly":"Disabled"}]}],"qosClass":"Burstable"}}config/pod/openshift-dns/logs/dns-default-xg7vh/kube-rbac-proxy_current.log0000640000000000000000000000156615114274105025552 0ustar00000000000000002025-12-04T11:38:33.791904182Z W1204 11:38:33.790711 1 deprecated.go:66] 2025-12-04T11:38:33.791904182Z ==== Removed Flag Warning ====================== 2025-12-04T11:38:33.791904182Z 2025-12-04T11:38:33.791904182Z logtostderr is removed in the k8s upstream and has no effect any more. 2025-12-04T11:38:33.791904182Z 2025-12-04T11:38:33.791904182Z =============================================== 2025-12-04T11:38:33.791904182Z 2025-12-04T11:38:33.791904182Z I1204 11:38:33.791406 1 kube-rbac-proxy.go:233] Valid token audiences: 2025-12-04T11:38:33.798419860Z I1204 11:38:33.797110 1 kube-rbac-proxy.go:347] Reading certificate files 2025-12-04T11:38:33.798419860Z I1204 11:38:33.797919 1 kube-rbac-proxy.go:395] Starting TCP socket on :9154 2025-12-04T11:38:33.801078206Z I1204 11:38:33.801023 1 kube-rbac-proxy.go:402] Listening securely on :9154 config/pod/openshift-dns/logs/dns-default-xg7vh/dns_current.log0000640000000000000000000001705415114274105023323 0ustar00000000000000002025-12-04T11:38:33.341585112Z .:5353 2025-12-04T11:38:33.341585112Z hostname.bind.:5353 2025-12-04T11:38:33.341886661Z [INFO] plugin/reload: Running configuration SHA512 = c40f1fac74a6633c6b1943fe251ad80adf3d5bd9b35c9e7d9b72bc260c5e2455f03e403e3b79d32f0936ff27e81ff6d07c68a95724b1c2c23510644372976718 2025-12-04T11:38:33.341886661Z CoreDNS-1.11.3 2025-12-04T11:38:33.341886661Z linux/amd64, go1.22.12 (Red Hat 1.22.12-3.el9_5) X:strictfipsruntime, 2025-12-04T11:50:44.387189245Z [INFO] 10.128.0.21:33619 - 10759 "AAAA IN thanos-querier.openshift-monitoring.svc. udp 68 false 1232" NXDOMAIN qr,rd,ra 57 0.001070982s 2025-12-04T11:50:44.387189245Z [INFO] 10.128.0.21:42542 - 37717 "A IN thanos-querier.openshift-monitoring.svc. udp 68 false 1232" NXDOMAIN qr,rd,ra 57 0.001222637s 2025-12-04T11:51:03.273431619Z [INFO] 10.128.0.21:59453 - 25055 "AAAA IN thanos-querier.openshift-monitoring.svc. udp 68 false 1232" NXDOMAIN qr,rd,ra 57 0.002940649s 2025-12-04T11:51:03.273593254Z [INFO] 10.128.0.21:41553 - 15946 "A IN thanos-querier.openshift-monitoring.svc. udp 68 false 1232" NXDOMAIN qr,rd,ra 57 0.003099073s 2025-12-04T11:51:25.359819887Z [INFO] 10.128.0.21:47912 - 43776 "AAAA IN thanos-querier.openshift-monitoring.svc. udp 68 false 1232" NXDOMAIN qr,rd,ra 57 0.002057662s 2025-12-04T11:51:25.359972881Z [INFO] 10.128.0.21:38540 - 40805 "A IN thanos-querier.openshift-monitoring.svc. udp 68 false 1232" NXDOMAIN qr,rd,ra 57 0.00231849s 2025-12-04T11:52:03.271996974Z [INFO] 10.128.0.21:49951 - 30003 "AAAA IN thanos-querier.openshift-monitoring.svc. udp 68 false 1232" NXDOMAIN qr,rd,ra 57 0.001487939s 2025-12-04T11:52:03.271996974Z [INFO] 10.128.0.21:40209 - 24567 "A IN thanos-querier.openshift-monitoring.svc. udp 68 false 1232" NXDOMAIN qr,rd,ra 57 0.001412986s 2025-12-04T11:53:03.273013913Z [INFO] 10.128.0.21:58185 - 28792 "AAAA IN thanos-querier.openshift-monitoring.svc. udp 68 false 1232" NXDOMAIN qr,rd,ra 57 0.001751386s 2025-12-04T11:53:03.273126287Z [INFO] 10.128.0.21:34216 - 42317 "A IN thanos-querier.openshift-monitoring.svc. udp 68 false 1232" NXDOMAIN qr,rd,ra 57 0.001799667s 2025-12-04T11:54:03.272304555Z [INFO] 10.128.0.21:52860 - 21577 "A IN thanos-querier.openshift-monitoring.svc. udp 68 false 1232" NXDOMAIN qr,rd,ra 57 0.001463597s 2025-12-04T11:54:03.272384137Z [INFO] 10.128.0.21:57794 - 60366 "AAAA IN thanos-querier.openshift-monitoring.svc. udp 68 false 1232" NXDOMAIN qr,rd,ra 57 0.00156458s 2025-12-04T11:54:09.216966788Z [INFO] 10.128.0.21:60273 - 45816 "A IN thanos-querier.openshift-monitoring.svc. udp 68 false 1232" NXDOMAIN qr,rd,ra 57 0.001117875s 2025-12-04T11:54:09.216966788Z [INFO] 10.128.0.21:44154 - 47599 "AAAA IN thanos-querier.openshift-monitoring.svc. udp 68 false 1232" NXDOMAIN qr,rd,ra 57 0.001200978s 2025-12-04T11:55:03.272292098Z [INFO] 10.128.0.21:56732 - 56028 "A IN thanos-querier.openshift-monitoring.svc. udp 68 false 1232" NXDOMAIN qr,rd,ra 57 0.001348112s 2025-12-04T11:55:03.272292098Z [INFO] 10.128.0.21:60057 - 21979 "AAAA IN thanos-querier.openshift-monitoring.svc. udp 68 false 1232" NXDOMAIN qr,rd,ra 57 0.001419603s 2025-12-04T12:01:09.530722540Z [INFO] 10.128.0.21:47576 - 32098 "AAAA IN thanos-querier.openshift-monitoring.svc. udp 68 false 1232" NXDOMAIN qr,rd,ra 57 0.002252861s 2025-12-04T12:01:09.530780162Z [INFO] 10.128.0.21:47131 - 43515 "A IN thanos-querier.openshift-monitoring.svc. udp 68 false 1232" NXDOMAIN qr,rd,ra 57 0.002700815s 2025-12-04T12:01:09.544138675Z [INFO] 10.128.0.21:44681 - 48323 "AAAA IN thanos-querier.openshift-monitoring.svc. udp 68 false 1232" NXDOMAIN qr,rd,ra 57 0.001539058s 2025-12-04T12:01:09.544409554Z [INFO] 10.128.0.21:51154 - 64640 "A IN thanos-querier.openshift-monitoring.svc. udp 68 false 1232" NXDOMAIN qr,rd,ra 57 0.00157962s 2025-12-04T12:01:09.562202228Z [INFO] 10.128.0.21:49685 - 29302 "A IN thanos-querier.openshift-monitoring.svc. udp 68 false 1232" NXDOMAIN qr,rd,ra 57 0.000394393s 2025-12-04T12:01:09.562202228Z [INFO] 10.128.0.21:41886 - 47164 "AAAA IN thanos-querier.openshift-monitoring.svc. udp 68 false 1232" NXDOMAIN qr,rd,ra 57 0.000489856s 2025-12-04T12:01:09.588616274Z [INFO] 10.128.0.21:54102 - 63246 "A IN thanos-querier.openshift-monitoring.svc. udp 68 false 1232" NXDOMAIN qr,rd,ra 57 0.000421394s 2025-12-04T12:01:09.588616274Z [INFO] 10.128.0.21:56768 - 54394 "AAAA IN thanos-querier.openshift-monitoring.svc. udp 68 false 1232" NXDOMAIN qr,rd,ra 57 0.000466854s 2025-12-04T12:01:09.634871809Z [INFO] 10.128.0.21:35930 - 13303 "A IN thanos-querier.openshift-monitoring.svc. udp 68 false 1232" NXDOMAIN qr,rd,ra 57 0.00062755s 2025-12-04T12:01:09.635058075Z [INFO] 10.128.0.21:48240 - 4163 "AAAA IN thanos-querier.openshift-monitoring.svc. udp 68 false 1232" NXDOMAIN qr,rd,ra 57 0.000856518s 2025-12-04T12:01:09.720597984Z [INFO] 10.128.0.21:40051 - 22412 "A IN thanos-querier.openshift-monitoring.svc. udp 68 false 1232" NXDOMAIN qr,rd,ra 57 0.000422352s 2025-12-04T12:01:09.720597984Z [INFO] 10.128.0.21:38458 - 41288 "AAAA IN thanos-querier.openshift-monitoring.svc. udp 68 false 1232" NXDOMAIN qr,rd,ra 57 0.000581748s 2025-12-04T12:01:09.887940515Z [INFO] 10.128.0.21:55645 - 19519 "A IN thanos-querier.openshift-monitoring.svc. udp 68 false 1232" NXDOMAIN qr,rd,ra 57 0.000434304s 2025-12-04T12:01:09.887940515Z [INFO] 10.128.0.21:43467 - 62784 "AAAA IN thanos-querier.openshift-monitoring.svc. udp 68 false 1232" NXDOMAIN qr,rd,ra 57 0.000459055s 2025-12-04T12:01:10.213686541Z [INFO] 10.128.0.21:48090 - 40320 "AAAA IN thanos-querier.openshift-monitoring.svc. udp 68 false 1232" NXDOMAIN qr,rd,ra 57 0.000489866s 2025-12-04T12:01:10.213686541Z [INFO] 10.128.0.21:36191 - 48436 "A IN thanos-querier.openshift-monitoring.svc. udp 68 false 1232" NXDOMAIN qr,rd,ra 57 0.000509956s 2025-12-04T12:01:10.859544705Z [INFO] 10.128.0.21:59334 - 54257 "A IN thanos-querier.openshift-monitoring.svc. udp 68 false 1232" NXDOMAIN qr,rd,ra 57 0.000401563s 2025-12-04T12:01:10.859544705Z [INFO] 10.128.0.21:50324 - 22224 "AAAA IN thanos-querier.openshift-monitoring.svc. udp 68 false 1232" NXDOMAIN qr,rd,ra 57 0.000459825s 2025-12-04T12:01:12.145764557Z [INFO] 10.128.0.21:34283 - 34245 "AAAA IN thanos-querier.openshift-monitoring.svc. udp 68 false 1232" NXDOMAIN qr,rd,ra 57 0.000651801s 2025-12-04T12:01:12.145764557Z [INFO] 10.128.0.21:58986 - 31064 "A IN thanos-querier.openshift-monitoring.svc. udp 68 false 1232" NXDOMAIN qr,rd,ra 57 0.000665821s 2025-12-04T12:01:14.713852645Z [INFO] 10.128.0.21:53970 - 64434 "AAAA IN thanos-querier.openshift-monitoring.svc. udp 68 false 1232" NXDOMAIN qr,rd,ra 57 0.000599759s 2025-12-04T12:01:14.713936027Z [INFO] 10.128.0.21:50066 - 50330 "A IN thanos-querier.openshift-monitoring.svc. udp 68 false 1232" NXDOMAIN qr,rd,ra 57 0.000706363s 2025-12-04T12:01:19.840047863Z [INFO] 10.128.0.21:51121 - 3247 "AAAA IN thanos-querier.openshift-monitoring.svc. udp 68 false 1232" NXDOMAIN qr,rd,ra 57 0.000721633s 2025-12-04T12:01:19.840106255Z [INFO] 10.128.0.21:53376 - 38967 "A IN thanos-querier.openshift-monitoring.svc. udp 68 false 1232" NXDOMAIN qr,rd,ra 57 0.000774764s 2025-12-04T12:01:30.090980615Z [INFO] 10.128.0.21:51886 - 10924 "AAAA IN thanos-querier.openshift-monitoring.svc. udp 68 false 1232" NXDOMAIN qr,rd,ra 57 0.00191898s 2025-12-04T12:01:30.091327316Z [INFO] 10.128.0.21:43440 - 12290 "A IN thanos-querier.openshift-monitoring.svc. udp 68 false 1232" NXDOMAIN qr,rd,ra 57 0.002285783s 2025-12-04T12:01:50.580172144Z [INFO] 10.128.0.21:46812 - 14399 "A IN thanos-querier.openshift-monitoring.svc. udp 68 false 1232" NXDOMAIN qr,rd,ra 57 0.001647522s 2025-12-04T12:01:50.580269327Z [INFO] 10.128.0.21:44292 - 4183 "AAAA IN thanos-querier.openshift-monitoring.svc. udp 68 false 1232" NXDOMAIN qr,rd,ra 57 0.001764966s config/pod/openshift-dns-operator/logs/dns-operator-7c56cf9b74-xz27r/kube-rbac-proxy_current.log0000640000000000000000000000156615114274105031144 0ustar00000000000000002025-12-04T11:38:20.165893562Z W1204 11:38:20.160699 1 deprecated.go:66] 2025-12-04T11:38:20.165893562Z ==== Removed Flag Warning ====================== 2025-12-04T11:38:20.165893562Z 2025-12-04T11:38:20.165893562Z logtostderr is removed in the k8s upstream and has no effect any more. 2025-12-04T11:38:20.165893562Z 2025-12-04T11:38:20.165893562Z =============================================== 2025-12-04T11:38:20.165893562Z 2025-12-04T11:38:20.165893562Z I1204 11:38:20.164778 1 kube-rbac-proxy.go:233] Valid token audiences: 2025-12-04T11:38:20.169199119Z I1204 11:38:20.166266 1 kube-rbac-proxy.go:347] Reading certificate files 2025-12-04T11:38:20.169199119Z I1204 11:38:20.166700 1 kube-rbac-proxy.go:395] Starting TCP socket on :9393 2025-12-04T11:38:20.169199119Z I1204 11:38:20.167227 1 kube-rbac-proxy.go:402] Listening securely on :9393 config/pod/openshift-dns-operator/logs/dns-operator-7c56cf9b74-xz27r/dns-operator_current.log0000640000000000000000000006176315114274105030554 0ustar00000000000000002025-12-04T11:38:19.756651312Z I1204 11:38:19.755311 1 simple_featuregate_reader.go:171] Starting feature-gate-detector 2025-12-04T11:38:19.766199369Z I1204 11:38:19.766082 1 event.go:377] Event(v1.ObjectReference{Kind:"Deployment", Namespace:"openshift-dns-operator", Name:"dns-operator", UID:"", APIVersion:"apps/v1", ResourceVersion:"", FieldPath:""}): type: 'Normal' reason: 'FeatureGatesInitialized' FeatureGates updated to featuregates.Features{Enabled:[]v1.FeatureGateName{"AWSEFSDriverVolumeMetrics", "AdminNetworkPolicy", "AlibabaPlatform", "AzureWorkloadIdentity", "BareMetalLoadBalancer", "BuildCSIVolumes", "ChunkSizeMiB", "CloudDualStackNodeIPs", "DisableKubeletCloudCredentialProviders", "GCPLabelsTags", "HardwareSpeed", "IngressControllerLBSubnetsAWS", "KMSv1", "ManagedBootImages", "ManagedBootImagesAWS", "MultiArchInstallAWS", "MultiArchInstallGCP", "NetworkDiagnosticsConfig", "NetworkLiveMigration", "NetworkSegmentation", "NewOLM", "NodeDisruptionPolicy", "OnClusterBuild", "PersistentIPsForVirtualization", "PrivateHostedZoneAWS", "SetEIPForNLBIngressController", "VSphereControlPlaneMachineSet", "VSphereDriverConfiguration", "VSphereMultiVCenters", "VSphereStaticIPs", "ValidatingAdmissionPolicy"}, Disabled:[]v1.FeatureGateName{"AWSClusterHostedDNS", "AdditionalRoutingCapabilities", "AutomatedEtcdBackup", "BootcNodeManagement", "CSIDriverSharedResource", "ClusterAPIInstall", "ClusterAPIInstallIBMCloud", "ClusterMonitoringConfig", "ConsolePluginContentSecurityPolicy", "DNSNameResolver", "DynamicResourceAllocation", "EtcdBackendQuota", "EventedPLEG", "Example", "ExternalOIDC", "ExternalOIDCWithUIDAndExtraClaimMappings", "GCPClusterHostedDNS", "GatewayAPI", "ImageStreamImportMode", "IngressControllerDynamicConfigurationManager", "InsightsConfig", "InsightsConfigAPI", "InsightsOnDemandDataGather", "InsightsRuntimeExtractor", "MachineAPIMigration", "MachineAPIOperatorDisableMachineHealthCheckController", "MachineAPIProviderOpenStack", "MachineConfigNodes", "MaxUnavailableStatefulSet", "MetricsCollectionProfiles", "MinimumKubeletVersion", "MixedCPUsAllocation", "MultiArchInstallAzure", "NodeSwap", "NutanixMultiSubnets", "OVNObservability", "OpenShiftPodSecurityAdmission", "PinnedImages", "PlatformOperators", "ProcMountType", "RouteAdvertisements", "RouteExternalCertificate", "ServiceAccountTokenNodeBinding", "SignatureStores", "SigstoreImageVerification", "StreamingCollectionEncodingToJSON", "StreamingCollectionEncodingToProtobuf", "TranslateStreamCloseWebsocketRequests", "UpgradeStatus", "UserNamespacesPodSecurityStandards", "UserNamespacesSupport", "VSphereMultiNetworks", "VolumeAttributesClass", "VolumeGroupSnapshot"}} 2025-12-04T11:38:19.766199369Z time="2025-12-04T11:38:19Z" level=info msg="FeatureGates initializedknownFeatures[AWSClusterHostedDNS AWSEFSDriverVolumeMetrics AdditionalRoutingCapabilities AdminNetworkPolicy AlibabaPlatform AutomatedEtcdBackup AzureWorkloadIdentity BareMetalLoadBalancer BootcNodeManagement BuildCSIVolumes CSIDriverSharedResource ChunkSizeMiB CloudDualStackNodeIPs ClusterAPIInstall ClusterAPIInstallIBMCloud ClusterMonitoringConfig ConsolePluginContentSecurityPolicy DNSNameResolver DisableKubeletCloudCredentialProviders DynamicResourceAllocation EtcdBackendQuota EventedPLEG Example ExternalOIDC ExternalOIDCWithUIDAndExtraClaimMappings GCPClusterHostedDNS GCPLabelsTags GatewayAPI HardwareSpeed ImageStreamImportMode IngressControllerDynamicConfigurationManager IngressControllerLBSubnetsAWS InsightsConfig InsightsConfigAPI InsightsOnDemandDataGather InsightsRuntimeExtractor KMSv1 MachineAPIMigration MachineAPIOperatorDisableMachineHealthCheckController MachineAPIProviderOpenStack MachineConfigNodes ManagedBootImages ManagedBootImagesAWS MaxUnavailableStatefulSet MetricsCollectionProfiles MinimumKubeletVersion MixedCPUsAllocation MultiArchInstallAWS MultiArchInstallAzure MultiArchInstallGCP NetworkDiagnosticsConfig NetworkLiveMigration NetworkSegmentation NewOLM NodeDisruptionPolicy NodeSwap NutanixMultiSubnets OVNObservability OnClusterBuild OpenShiftPodSecurityAdmission PersistentIPsForVirtualization PinnedImages PlatformOperators PrivateHostedZoneAWS ProcMountType RouteAdvertisements RouteExternalCertificate ServiceAccountTokenNodeBinding SetEIPForNLBIngressController SignatureStores SigstoreImageVerification StreamingCollectionEncodingToJSON StreamingCollectionEncodingToProtobuf TranslateStreamCloseWebsocketRequests UpgradeStatus UserNamespacesPodSecurityStandards UserNamespacesSupport VSphereControlPlaneMachineSet VSphereDriverConfiguration VSphereMultiNetworks VSphereMultiVCenters VSphereStaticIPs ValidatingAdmissionPolicy VolumeAttributesClass VolumeGroupSnapshot]" 2025-12-04T11:38:19.766715935Z 2025-12-04T11:38:19Z INFO controller-runtime.metrics Starting metrics server 2025-12-04T11:38:19.767308554Z 2025-12-04T11:38:19Z INFO controller-runtime.metrics Serving metrics server {"bindAddress": "127.0.0.1:60000", "secure": false} 2025-12-04T11:38:19.767308554Z 2025-12-04T11:38:19Z INFO Starting EventSource {"controller": "status_controller", "source": "kind source: *v1.DNS"} 2025-12-04T11:38:19.767373096Z 2025-12-04T11:38:19Z INFO Starting EventSource {"controller": "dns_controller", "source": "kind source: *v1.DNS"} 2025-12-04T11:38:19.767393777Z 2025-12-04T11:38:19Z INFO Starting EventSource {"controller": "dns_controller", "source": "kind source: *v1.DaemonSet"} 2025-12-04T11:38:19.767393777Z 2025-12-04T11:38:19Z INFO Starting EventSource {"controller": "dns_controller", "source": "kind source: *v1.Service"} 2025-12-04T11:38:19.767393777Z 2025-12-04T11:38:19Z INFO Starting EventSource {"controller": "dns_controller", "source": "kind source: *v1.ConfigMap"} 2025-12-04T11:38:19.767429998Z 2025-12-04T11:38:19Z INFO Starting EventSource {"controller": "dns_controller", "source": "kind source: *v1.ConfigMap"} 2025-12-04T11:38:19.767429998Z 2025-12-04T11:38:19Z INFO Starting EventSource {"controller": "status_controller", "source": "kind source: *v1.DaemonSet"} 2025-12-04T11:38:19.767429998Z 2025-12-04T11:38:19Z INFO Starting EventSource {"controller": "dns_controller", "source": "kind source: *v1.Node"} 2025-12-04T11:38:19.767439628Z 2025-12-04T11:38:19Z INFO Starting Controller {"controller": "dns_controller"} 2025-12-04T11:38:19.767439628Z 2025-12-04T11:38:19Z INFO Starting EventSource {"controller": "status_controller", "source": "kind source: *v1.ClusterOperator"} 2025-12-04T11:38:19.767447588Z 2025-12-04T11:38:19Z INFO Starting Controller {"controller": "status_controller"} 2025-12-04T11:38:19.851470186Z time="2025-12-04T11:38:19Z" level=info msg="created default dns: default" 2025-12-04T11:38:20.122635323Z 2025-12-04T11:38:20Z INFO Starting workers {"controller": "dns_controller", "worker count": 1} 2025-12-04T11:38:20.122771387Z time="2025-12-04T11:38:20Z" level=info msg="reconciling request: /default" 2025-12-04T11:38:20.133519122Z 2025-12-04T11:38:20Z INFO Starting workers {"controller": "status_controller", "worker count": 1} 2025-12-04T11:38:20.151971525Z time="2025-12-04T11:38:20Z" level=info msg="created dns namespace: openshift-dns" 2025-12-04T11:38:20.214415470Z time="2025-12-04T11:38:20Z" level=info msg="created dns cluster role: /openshift-dns" 2025-12-04T11:38:20.311551899Z time="2025-12-04T11:38:20Z" level=info msg="created dns cluster role binding: openshift-dns" 2025-12-04T11:38:20.494021758Z time="2025-12-04T11:38:20Z" level=info msg="created dns service account: openshift-dns/dns" 2025-12-04T11:38:20.511632203Z time="2025-12-04T11:38:20Z" level=info msg="created serviceaccount openshift-dns/node-resolver" 2025-12-04T11:38:20.534877750Z time="2025-12-04T11:38:20Z" level=info msg="enforced finalizer for dns: default" 2025-12-04T11:38:20.599030710Z time="2025-12-04T11:38:20Z" level=info msg="created dns daemonset: openshift-dns/dns-default" 2025-12-04T11:38:20.785359262Z time="2025-12-04T11:38:20Z" level=info msg="created configmap: dns-default" 2025-12-04T11:38:20.948042206Z time="2025-12-04T11:38:20Z" level=info msg="created dns service: openshift-dns/dns-default" 2025-12-04T11:38:20.978288538Z time="2025-12-04T11:38:20Z" level=info msg="created dns metrics cluster role dns-monitoring" 2025-12-04T11:38:21.115858964Z time="2025-12-04T11:38:21Z" level=info msg="created dns metrics cluster role binding dns-monitoring" 2025-12-04T11:38:21.237482930Z time="2025-12-04T11:38:21Z" level=info msg="created dns metrics role openshift-dns/prometheus-k8s" 2025-12-04T11:38:21.293395425Z time="2025-12-04T11:38:21Z" level=info msg="created dns metrics role binding openshift-dns/prometheus-k8s" 2025-12-04T11:38:21.307060354Z time="2025-12-04T11:38:21Z" level=info msg="created servicemonitor openshift-dns/dns-default" 2025-12-04T11:38:21.425771406Z time="2025-12-04T11:38:21Z" level=info msg="created node resolver daemonset: openshift-dns/node-resolver" 2025-12-04T11:38:21.452158843Z time="2025-12-04T11:38:21Z" level=info msg="updated DNS default status: old: v1.DNSStatus{ClusterIP:\"\", ClusterDomain:\"\", Conditions:[]v1.OperatorCondition(nil)}, new: v1.DNSStatus{ClusterIP:\"172.30.0.10\", ClusterDomain:\"cluster.local\", Conditions:[]v1.OperatorCondition{v1.OperatorCondition{Type:\"Degraded\", Status:\"True\", LastTransitionTime:time.Date(2025, time.December, 4, 11, 38, 21, 0, time.Local), Reason:\"NoDNSPodsDesired\", Message:\"No DNS pods are desired; this could mean all nodes are tainted or unschedulable.\"}, v1.OperatorCondition{Type:\"Progressing\", Status:\"False\", LastTransitionTime:time.Date(2025, time.December, 4, 11, 38, 21, 0, time.Local), Reason:\"AsExpected\", Message:\"All DNS and node-resolver pods are available, and the DNS service has a cluster IP address.\"}, v1.OperatorCondition{Type:\"Available\", Status:\"False\", LastTransitionTime:time.Date(2025, time.December, 4, 11, 38, 21, 0, time.Local), Reason:\"NoDaemonSetPods\", Message:\"The DNS daemonset has no pods available.\"}, v1.OperatorCondition{Type:\"Upgradeable\", Status:\"True\", LastTransitionTime:time.Date(2025, time.December, 4, 11, 38, 21, 0, time.Local), Reason:\"AsExpected\", Message:\"DNS Operator can be upgraded\"}}}" 2025-12-04T11:38:21.470318096Z time="2025-12-04T11:38:21Z" level=info msg="created external name service default/openshift" 2025-12-04T11:38:21.470399429Z time="2025-12-04T11:38:21Z" level=info msg="reconciling request: /default" 2025-12-04T11:38:21.646709520Z time="2025-12-04T11:38:21Z" level=info msg="updated DNS default status: old: v1.DNSStatus{ClusterIP:\"172.30.0.10\", ClusterDomain:\"cluster.local\", Conditions:[]v1.OperatorCondition{v1.OperatorCondition{Type:\"Degraded\", Status:\"True\", LastTransitionTime:time.Date(2025, time.December, 4, 11, 38, 21, 0, time.Local), Reason:\"NoDNSPodsDesired\", Message:\"No DNS pods are desired; this could mean all nodes are tainted or unschedulable.\"}, v1.OperatorCondition{Type:\"Progressing\", Status:\"False\", LastTransitionTime:time.Date(2025, time.December, 4, 11, 38, 21, 0, time.Local), Reason:\"AsExpected\", Message:\"All DNS and node-resolver pods are available, and the DNS service has a cluster IP address.\"}, v1.OperatorCondition{Type:\"Available\", Status:\"False\", LastTransitionTime:time.Date(2025, time.December, 4, 11, 38, 21, 0, time.Local), Reason:\"NoDaemonSetPods\", Message:\"The DNS daemonset has no pods available.\"}, v1.OperatorCondition{Type:\"Upgradeable\", Status:\"True\", LastTransitionTime:time.Date(2025, time.December, 4, 11, 38, 21, 0, time.Local), Reason:\"AsExpected\", Message:\"DNS Operator can be upgraded\"}}}, new: v1.DNSStatus{ClusterIP:\"172.30.0.10\", ClusterDomain:\"cluster.local\", Conditions:[]v1.OperatorCondition{v1.OperatorCondition{Type:\"Degraded\", Status:\"True\", LastTransitionTime:time.Date(2025, time.December, 4, 11, 38, 21, 0, time.Local), Reason:\"NoDNSPodsAvailable\", Message:\"No DNS pods are available.\"}, v1.OperatorCondition{Type:\"Progressing\", Status:\"False\", LastTransitionTime:time.Date(2025, time.December, 4, 11, 38, 21, 0, time.Local), Reason:\"AsExpected\", Message:\"All DNS and node-resolver pods are available, and the DNS service has a cluster IP address.\"}, v1.OperatorCondition{Type:\"Available\", Status:\"False\", LastTransitionTime:time.Date(2025, time.December, 4, 11, 38, 21, 0, time.Local), Reason:\"NoDaemonSetPods\", Message:\"The DNS daemonset has no pods available.\"}, v1.OperatorCondition{Type:\"Upgradeable\", Status:\"True\", LastTransitionTime:time.Date(2025, time.December, 4, 11, 38, 21, 0, time.Local), Reason:\"AsExpected\", Message:\"DNS Operator can be upgraded\"}}}" 2025-12-04T11:38:21.650058588Z time="2025-12-04T11:38:21Z" level=info msg="reconciling request: /default" 2025-12-04T11:38:33.538313419Z time="2025-12-04T11:38:33Z" level=info msg="reconciling request: /default" 2025-12-04T11:38:46.128946184Z time="2025-12-04T11:38:46Z" level=info msg="reconciling request: /default" 2025-12-04T11:39:01.650954682Z time="2025-12-04T11:39:01Z" level=info msg="reconciling request: /default" 2025-12-04T11:40:01.654829329Z time="2025-12-04T11:40:01Z" level=error msg="failed to reconcile request /default: failed to get dns /default: the server was unable to return a response in the time allotted, but may still be processing the request (get dnses.operator.openshift.io default)" 2025-12-04T11:40:01.654925882Z 2025-12-04T11:40:01Z ERROR Reconciler error {"controller": "dns_controller", "object": {"name":"default"}, "namespace": "", "name": "default", "reconcileID": "1d10d7c7-b867-4382-a4c0-1dce731b8c0e", "error": "failed to get dns /default: the server was unable to return a response in the time allotted, but may still be processing the request (get dnses.operator.openshift.io default)", "errorCauses": [{"error": "failed to get dns /default: the server was unable to return a response in the time allotted, but may still be processing the request (get dnses.operator.openshift.io default)"}]} 2025-12-04T11:40:01.654925882Z sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).reconcileHandler 2025-12-04T11:40:01.654925882Z /dns-operator/vendor/sigs.k8s.io/controller-runtime/pkg/internal/controller/controller.go:324 2025-12-04T11:40:01.654925882Z sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).processNextWorkItem 2025-12-04T11:40:01.654925882Z /dns-operator/vendor/sigs.k8s.io/controller-runtime/pkg/internal/controller/controller.go:261 2025-12-04T11:40:01.654925882Z sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Start.func2.2 2025-12-04T11:40:01.654925882Z /dns-operator/vendor/sigs.k8s.io/controller-runtime/pkg/internal/controller/controller.go:222 2025-12-04T11:40:01.660294140Z time="2025-12-04T11:40:01Z" level=info msg="reconciling request: /default" 2025-12-04T11:40:19.854298390Z time="2025-12-04T11:40:19Z" level=error msg="failed to ensure default dns the server was unable to return a response in the time allotted, but may still be processing the request (get dnses.operator.openshift.io default)" 2025-12-04T11:41:01.664185154Z time="2025-12-04T11:41:01Z" level=error msg="failed to reconcile request /default: failed to get dns /default: the server was unable to return a response in the time allotted, but may still be processing the request (get dnses.operator.openshift.io default)" 2025-12-04T11:41:01.664499164Z 2025-12-04T11:41:01Z ERROR Reconciler error {"controller": "dns_controller", "object": {"name":"default"}, "namespace": "", "name": "default", "reconcileID": "17e717eb-44c6-4bee-a65b-edecc597aa1b", "error": "failed to get dns /default: the server was unable to return a response in the time allotted, but may still be processing the request (get dnses.operator.openshift.io default)", "errorCauses": [{"error": "failed to get dns /default: the server was unable to return a response in the time allotted, but may still be processing the request (get dnses.operator.openshift.io default)"}]} 2025-12-04T11:41:01.664499164Z sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).reconcileHandler 2025-12-04T11:41:01.664499164Z /dns-operator/vendor/sigs.k8s.io/controller-runtime/pkg/internal/controller/controller.go:324 2025-12-04T11:41:01.664499164Z sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).processNextWorkItem 2025-12-04T11:41:01.664499164Z /dns-operator/vendor/sigs.k8s.io/controller-runtime/pkg/internal/controller/controller.go:261 2025-12-04T11:41:01.664499164Z sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Start.func2.2 2025-12-04T11:41:01.664499164Z /dns-operator/vendor/sigs.k8s.io/controller-runtime/pkg/internal/controller/controller.go:222 2025-12-04T11:41:01.674945961Z time="2025-12-04T11:41:01Z" level=info msg="reconciling request: /default" 2025-12-04T11:42:01.678619230Z time="2025-12-04T11:42:01Z" level=error msg="failed to reconcile request /default: failed to get dns /default: the server was unable to return a response in the time allotted, but may still be processing the request (get dnses.operator.openshift.io default)" 2025-12-04T11:42:01.678688233Z 2025-12-04T11:42:01Z ERROR Reconciler error {"controller": "dns_controller", "object": {"name":"default"}, "namespace": "", "name": "default", "reconcileID": "2be87eaf-43e3-4655-9a59-9d9bb100dc50", "error": "failed to get dns /default: the server was unable to return a response in the time allotted, but may still be processing the request (get dnses.operator.openshift.io default)", "errorCauses": [{"error": "failed to get dns /default: the server was unable to return a response in the time allotted, but may still be processing the request (get dnses.operator.openshift.io default)"}]} 2025-12-04T11:42:01.678688233Z sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).reconcileHandler 2025-12-04T11:42:01.678688233Z /dns-operator/vendor/sigs.k8s.io/controller-runtime/pkg/internal/controller/controller.go:324 2025-12-04T11:42:01.678688233Z sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).processNextWorkItem 2025-12-04T11:42:01.678688233Z /dns-operator/vendor/sigs.k8s.io/controller-runtime/pkg/internal/controller/controller.go:261 2025-12-04T11:42:01.678688233Z sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Start.func2.2 2025-12-04T11:42:01.678688233Z /dns-operator/vendor/sigs.k8s.io/controller-runtime/pkg/internal/controller/controller.go:222 2025-12-04T11:42:01.699267386Z time="2025-12-04T11:42:01Z" level=info msg="reconciling request: /default" 2025-12-04T11:42:19.856595400Z time="2025-12-04T11:42:19Z" level=error msg="failed to ensure default dns the server was unable to return a response in the time allotted, but may still be processing the request (get dnses.operator.openshift.io default)" 2025-12-04T11:43:01.703223632Z time="2025-12-04T11:43:01Z" level=error msg="failed to reconcile request /default: failed to get dns /default: the server was unable to return a response in the time allotted, but may still be processing the request (get dnses.operator.openshift.io default)" 2025-12-04T11:43:01.703350127Z 2025-12-04T11:43:01Z ERROR Reconciler error {"controller": "dns_controller", "object": {"name":"default"}, "namespace": "", "name": "default", "reconcileID": "33163fab-0b6a-4a1d-9e48-a1c682b85160", "error": "failed to get dns /default: the server was unable to return a response in the time allotted, but may still be processing the request (get dnses.operator.openshift.io default)", "errorCauses": [{"error": "failed to get dns /default: the server was unable to return a response in the time allotted, but may still be processing the request (get dnses.operator.openshift.io default)"}]} 2025-12-04T11:43:01.703350127Z sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).reconcileHandler 2025-12-04T11:43:01.703350127Z /dns-operator/vendor/sigs.k8s.io/controller-runtime/pkg/internal/controller/controller.go:324 2025-12-04T11:43:01.703350127Z sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).processNextWorkItem 2025-12-04T11:43:01.703350127Z /dns-operator/vendor/sigs.k8s.io/controller-runtime/pkg/internal/controller/controller.go:261 2025-12-04T11:43:01.703350127Z sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Start.func2.2 2025-12-04T11:43:01.703350127Z /dns-operator/vendor/sigs.k8s.io/controller-runtime/pkg/internal/controller/controller.go:222 2025-12-04T11:43:01.743942567Z time="2025-12-04T11:43:01Z" level=info msg="reconciling request: /default" 2025-12-04T11:44:01.746553397Z time="2025-12-04T11:44:01Z" level=error msg="failed to reconcile request /default: failed to get dns /default: the server was unable to return a response in the time allotted, but may still be processing the request (get dnses.operator.openshift.io default)" 2025-12-04T11:44:01.746655191Z 2025-12-04T11:44:01Z ERROR Reconciler error {"controller": "dns_controller", "object": {"name":"default"}, "namespace": "", "name": "default", "reconcileID": "cefd4d73-7a29-4227-9f06-38541b322ed0", "error": "failed to get dns /default: the server was unable to return a response in the time allotted, but may still be processing the request (get dnses.operator.openshift.io default)", "errorCauses": [{"error": "failed to get dns /default: the server was unable to return a response in the time allotted, but may still be processing the request (get dnses.operator.openshift.io default)"}]} 2025-12-04T11:44:01.746655191Z sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).reconcileHandler 2025-12-04T11:44:01.746655191Z /dns-operator/vendor/sigs.k8s.io/controller-runtime/pkg/internal/controller/controller.go:324 2025-12-04T11:44:01.746655191Z sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).processNextWorkItem 2025-12-04T11:44:01.746655191Z /dns-operator/vendor/sigs.k8s.io/controller-runtime/pkg/internal/controller/controller.go:261 2025-12-04T11:44:01.746655191Z sigs.k8s.io/controller-runtime/pkg/internal/controller.(*Controller).Start.func2.2 2025-12-04T11:44:01.746655191Z /dns-operator/vendor/sigs.k8s.io/controller-runtime/pkg/internal/controller/controller.go:222 2025-12-04T11:44:01.827744044Z time="2025-12-04T11:44:01Z" level=info msg="reconciling request: /default" 2025-12-04T11:45:38.777873922Z time="2025-12-04T11:45:38Z" level=info msg="updated DNS default status: old: v1.DNSStatus{ClusterIP:\"172.30.0.10\", ClusterDomain:\"cluster.local\", Conditions:[]v1.OperatorCondition{v1.OperatorCondition{Type:\"Degraded\", Status:\"True\", LastTransitionTime:time.Date(2025, time.December, 4, 11, 38, 21, 0, time.Local), Reason:\"NoDNSPodsAvailable\", Message:\"No DNS pods are available.\"}, v1.OperatorCondition{Type:\"Progressing\", Status:\"False\", LastTransitionTime:time.Date(2025, time.December, 4, 11, 38, 21, 0, time.Local), Reason:\"AsExpected\", Message:\"All DNS and node-resolver pods are available, and the DNS service has a cluster IP address.\"}, v1.OperatorCondition{Type:\"Available\", Status:\"False\", LastTransitionTime:time.Date(2025, time.December, 4, 11, 38, 21, 0, time.Local), Reason:\"NoDaemonSetPods\", Message:\"The DNS daemonset has no pods available.\"}, v1.OperatorCondition{Type:\"Upgradeable\", Status:\"True\", LastTransitionTime:time.Date(2025, time.December, 4, 11, 38, 21, 0, time.Local), Reason:\"AsExpected\", Message:\"DNS Operator can be upgraded\"}}}, new: v1.DNSStatus{ClusterIP:\"172.30.0.10\", ClusterDomain:\"cluster.local\", Conditions:[]v1.OperatorCondition{v1.OperatorCondition{Type:\"Degraded\", Status:\"True\", LastTransitionTime:time.Date(2025, time.December, 4, 11, 38, 21, 0, time.Local), Reason:\"NoDNSPodsAvailable\", Message:\"No DNS pods are available.\"}, v1.OperatorCondition{Type:\"Progressing\", Status:\"True\", LastTransitionTime:time.Date(2025, time.December, 4, 11, 45, 38, 0, time.Local), Reason:\"Reconciling\", Message:\"Have 0 available DNS pods, want 1.\"}, v1.OperatorCondition{Type:\"Available\", Status:\"False\", LastTransitionTime:time.Date(2025, time.December, 4, 11, 38, 21, 0, time.Local), Reason:\"NoDaemonSetPods\", Message:\"The DNS daemonset has no pods available.\"}, v1.OperatorCondition{Type:\"Upgradeable\", Status:\"True\", LastTransitionTime:time.Date(2025, time.December, 4, 11, 38, 21, 0, time.Local), Reason:\"AsExpected\", Message:\"DNS Operator can be upgraded\"}}}" 2025-12-04T11:45:38.782603846Z time="2025-12-04T11:45:38Z" level=info msg="reconciling request: /default" 2025-12-04T11:52:14.142656686Z time="2025-12-04T11:52:14Z" level=error msg="failed to ensure default dns the server was unable to return a response in the time allotted, but may still be processing the request (get dnses.operator.openshift.io default)" 2025-12-04T11:54:14.145012871Z time="2025-12-04T11:54:14Z" level=error msg="failed to ensure default dns the server was unable to return a response in the time allotted, but may still be processing the request (get dnses.operator.openshift.io default)" 2025-12-04T11:59:47.960861241Z time="2025-12-04T11:59:47Z" level=info msg="reconciling request: /default" 2025-12-04T11:59:51.108320334Z time="2025-12-04T11:59:51Z" level=info msg="reconciling request: /default" 2025-12-04T11:59:53.316248071Z time="2025-12-04T11:59:53Z" level=info msg="reconciling request: /default" 2025-12-04T11:59:53.926165818Z time="2025-12-04T11:59:53Z" level=info msg="reconciling request: /default" 2025-12-04T11:59:54.437032957Z time="2025-12-04T11:59:54Z" level=info msg="reconciling request: /default" 2025-12-04T11:59:58.097970703Z time="2025-12-04T11:59:58Z" level=info msg="reconciling request: /default" config/pod/openshift-oauth-apiserver/logs/apiserver-7467446c4b-dlj7g/oauth-apiserver_current.log0000640000000000000000000010752715114274105031206 0ustar00000000000000002025-12-04T11:57:32.183168975Z E1204 11:57:32.183148 1 authentication.go:73] "Unable to authenticate the request" err="[x509: certificate signed by unknown authority, verifying certificate SN=4812284156861786932, SKID=03:5A:81:6F:76:6B:16:1A:1E:F5:A6:47:1C:D2:81:87:CA:A1:6A:E6, AKID=BC:42:B1:AC:E1:64:5A:31:96:B1:4D:DA:91:1D:5C:CE:CA:10:21:26 failed: x509: certificate signed by unknown authority]" 2025-12-04T11:57:32.183223086Z E1204 11:57:32.183061 1 authentication.go:73] "Unable to authenticate the request" err="[x509: certificate signed by unknown authority, verifying certificate SN=4812284156861786932, SKID=03:5A:81:6F:76:6B:16:1A:1E:F5:A6:47:1C:D2:81:87:CA:A1:6A:E6, AKID=BC:42:B1:AC:E1:64:5A:31:96:B1:4D:DA:91:1D:5C:CE:CA:10:21:26 failed: x509: certificate signed by unknown authority]" 2025-12-04T11:57:32.183404781Z E1204 11:57:32.183357 1 authentication.go:73] "Unable to authenticate the request" err="[x509: certificate signed by unknown authority, verifying certificate SN=4812284156861786932, SKID=03:5A:81:6F:76:6B:16:1A:1E:F5:A6:47:1C:D2:81:87:CA:A1:6A:E6, AKID=BC:42:B1:AC:E1:64:5A:31:96:B1:4D:DA:91:1D:5C:CE:CA:10:21:26 failed: x509: certificate signed by unknown authority]" 2025-12-04T11:57:56.220468396Z E1204 11:57:56.220206 1 authentication.go:73] "Unable to authenticate the request" err="[x509: certificate signed by unknown authority, verifying certificate SN=4812284156861786932, SKID=03:5A:81:6F:76:6B:16:1A:1E:F5:A6:47:1C:D2:81:87:CA:A1:6A:E6, AKID=BC:42:B1:AC:E1:64:5A:31:96:B1:4D:DA:91:1D:5C:CE:CA:10:21:26 failed: x509: certificate signed by unknown authority]" 2025-12-04T11:57:56.220468396Z E1204 11:57:56.220291 1 authentication.go:73] "Unable to authenticate the request" err="[x509: certificate signed by unknown authority, verifying certificate SN=4812284156861786932, SKID=03:5A:81:6F:76:6B:16:1A:1E:F5:A6:47:1C:D2:81:87:CA:A1:6A:E6, AKID=BC:42:B1:AC:E1:64:5A:31:96:B1:4D:DA:91:1D:5C:CE:CA:10:21:26 failed: x509: certificate signed by unknown authority]" 2025-12-04T11:57:56.220566669Z E1204 11:57:56.220540 1 authentication.go:73] "Unable to authenticate the request" err="[x509: certificate signed by unknown authority, verifying certificate SN=4812284156861786932, SKID=03:5A:81:6F:76:6B:16:1A:1E:F5:A6:47:1C:D2:81:87:CA:A1:6A:E6, AKID=BC:42:B1:AC:E1:64:5A:31:96:B1:4D:DA:91:1D:5C:CE:CA:10:21:26 failed: x509: certificate signed by unknown authority]" 2025-12-04T11:57:56.220838006Z E1204 11:57:56.220795 1 authentication.go:73] "Unable to authenticate the request" err="[x509: certificate signed by unknown authority, verifying certificate SN=4812284156861786932, SKID=03:5A:81:6F:76:6B:16:1A:1E:F5:A6:47:1C:D2:81:87:CA:A1:6A:E6, AKID=BC:42:B1:AC:E1:64:5A:31:96:B1:4D:DA:91:1D:5C:CE:CA:10:21:26 failed: x509: certificate signed by unknown authority]" 2025-12-04T11:57:56.221117504Z E1204 11:57:56.221088 1 authentication.go:73] "Unable to authenticate the request" err="[x509: certificate signed by unknown authority, verifying certificate SN=4812284156861786932, SKID=03:5A:81:6F:76:6B:16:1A:1E:F5:A6:47:1C:D2:81:87:CA:A1:6A:E6, AKID=BC:42:B1:AC:E1:64:5A:31:96:B1:4D:DA:91:1D:5C:CE:CA:10:21:26 failed: x509: certificate signed by unknown authority]" 2025-12-04T11:57:56.221216386Z E1204 11:57:56.221187 1 authentication.go:73] "Unable to authenticate the request" err="[x509: certificate signed by unknown authority, verifying certificate SN=4812284156861786932, SKID=03:5A:81:6F:76:6B:16:1A:1E:F5:A6:47:1C:D2:81:87:CA:A1:6A:E6, AKID=BC:42:B1:AC:E1:64:5A:31:96:B1:4D:DA:91:1D:5C:CE:CA:10:21:26 failed: x509: certificate signed by unknown authority]" 2025-12-04T11:57:56.221314669Z E1204 11:57:56.221290 1 authentication.go:73] "Unable to authenticate the request" err="[x509: certificate signed by unknown authority, verifying certificate SN=4812284156861786932, SKID=03:5A:81:6F:76:6B:16:1A:1E:F5:A6:47:1C:D2:81:87:CA:A1:6A:E6, AKID=BC:42:B1:AC:E1:64:5A:31:96:B1:4D:DA:91:1D:5C:CE:CA:10:21:26 failed: x509: certificate signed by unknown authority]" 2025-12-04T11:57:56.221460373Z E1204 11:57:56.221421 1 authentication.go:73] "Unable to authenticate the request" err="[x509: certificate signed by unknown authority, verifying certificate SN=4812284156861786932, SKID=03:5A:81:6F:76:6B:16:1A:1E:F5:A6:47:1C:D2:81:87:CA:A1:6A:E6, AKID=BC:42:B1:AC:E1:64:5A:31:96:B1:4D:DA:91:1D:5C:CE:CA:10:21:26 failed: x509: certificate signed by unknown authority]" 2025-12-04T11:57:56.221756161Z E1204 11:57:56.221421 1 authentication.go:73] "Unable to authenticate the request" err="[x509: certificate signed by unknown authority, verifying certificate SN=4812284156861786932, SKID=03:5A:81:6F:76:6B:16:1A:1E:F5:A6:47:1C:D2:81:87:CA:A1:6A:E6, AKID=BC:42:B1:AC:E1:64:5A:31:96:B1:4D:DA:91:1D:5C:CE:CA:10:21:26 failed: x509: certificate signed by unknown authority]" 2025-12-04T11:57:56.221947616Z E1204 11:57:56.221428 1 authentication.go:73] "Unable to authenticate the request" err="[x509: certificate signed by unknown authority, verifying certificate SN=4812284156861786932, SKID=03:5A:81:6F:76:6B:16:1A:1E:F5:A6:47:1C:D2:81:87:CA:A1:6A:E6, AKID=BC:42:B1:AC:E1:64:5A:31:96:B1:4D:DA:91:1D:5C:CE:CA:10:21:26 failed: x509: certificate signed by unknown authority]" 2025-12-04T11:58:02.152963067Z E1204 11:58:02.152400 1 authentication.go:73] "Unable to authenticate the request" err="[x509: certificate signed by unknown authority, verifying certificate SN=4812284156861786932, SKID=03:5A:81:6F:76:6B:16:1A:1E:F5:A6:47:1C:D2:81:87:CA:A1:6A:E6, AKID=BC:42:B1:AC:E1:64:5A:31:96:B1:4D:DA:91:1D:5C:CE:CA:10:21:26 failed: x509: certificate signed by unknown authority]" 2025-12-04T11:58:02.152963067Z E1204 11:58:02.152429 1 authentication.go:73] "Unable to authenticate the request" err="[x509: certificate signed by unknown authority, verifying certificate SN=4812284156861786932, SKID=03:5A:81:6F:76:6B:16:1A:1E:F5:A6:47:1C:D2:81:87:CA:A1:6A:E6, AKID=BC:42:B1:AC:E1:64:5A:31:96:B1:4D:DA:91:1D:5C:CE:CA:10:21:26 failed: x509: certificate signed by unknown authority]" 2025-12-04T11:58:02.152963067Z E1204 11:58:02.152400 1 authentication.go:73] "Unable to authenticate the request" err="[x509: certificate signed by unknown authority, verifying certificate SN=4812284156861786932, SKID=03:5A:81:6F:76:6B:16:1A:1E:F5:A6:47:1C:D2:81:87:CA:A1:6A:E6, AKID=BC:42:B1:AC:E1:64:5A:31:96:B1:4D:DA:91:1D:5C:CE:CA:10:21:26 failed: x509: certificate signed by unknown authority]" 2025-12-04T11:58:02.152963067Z E1204 11:58:02.152575 1 authentication.go:73] "Unable to authenticate the request" err="[x509: certificate signed by unknown authority, verifying certificate SN=4812284156861786932, SKID=03:5A:81:6F:76:6B:16:1A:1E:F5:A6:47:1C:D2:81:87:CA:A1:6A:E6, AKID=BC:42:B1:AC:E1:64:5A:31:96:B1:4D:DA:91:1D:5C:CE:CA:10:21:26 failed: x509: certificate signed by unknown authority]" 2025-12-04T11:58:02.152963067Z E1204 11:58:02.152429 1 authentication.go:73] "Unable to authenticate the request" err="[x509: certificate signed by unknown authority, verifying certificate SN=4812284156861786932, SKID=03:5A:81:6F:76:6B:16:1A:1E:F5:A6:47:1C:D2:81:87:CA:A1:6A:E6, AKID=BC:42:B1:AC:E1:64:5A:31:96:B1:4D:DA:91:1D:5C:CE:CA:10:21:26 failed: x509: certificate signed by unknown authority]" 2025-12-04T11:58:02.191212199Z E1204 11:58:02.191121 1 authentication.go:73] "Unable to authenticate the request" err="[x509: certificate signed by unknown authority, verifying certificate SN=4812284156861786932, SKID=03:5A:81:6F:76:6B:16:1A:1E:F5:A6:47:1C:D2:81:87:CA:A1:6A:E6, AKID=BC:42:B1:AC:E1:64:5A:31:96:B1:4D:DA:91:1D:5C:CE:CA:10:21:26 failed: x509: certificate signed by unknown authority]" 2025-12-04T11:58:02.191212199Z E1204 11:58:02.191174 1 authentication.go:73] "Unable to authenticate the request" err="[x509: certificate signed by unknown authority, verifying certificate SN=4812284156861786932, SKID=03:5A:81:6F:76:6B:16:1A:1E:F5:A6:47:1C:D2:81:87:CA:A1:6A:E6, AKID=BC:42:B1:AC:E1:64:5A:31:96:B1:4D:DA:91:1D:5C:CE:CA:10:21:26 failed: x509: certificate signed by unknown authority]" 2025-12-04T11:58:02.191212199Z E1204 11:58:02.191181 1 authentication.go:73] "Unable to authenticate the request" err="[x509: certificate signed by unknown authority, verifying certificate SN=4812284156861786932, SKID=03:5A:81:6F:76:6B:16:1A:1E:F5:A6:47:1C:D2:81:87:CA:A1:6A:E6, AKID=BC:42:B1:AC:E1:64:5A:31:96:B1:4D:DA:91:1D:5C:CE:CA:10:21:26 failed: x509: certificate signed by unknown authority]" 2025-12-04T11:58:02.191297841Z E1204 11:58:02.191283 1 authentication.go:73] "Unable to authenticate the request" err="[x509: certificate signed by unknown authority, verifying certificate SN=4812284156861786932, SKID=03:5A:81:6F:76:6B:16:1A:1E:F5:A6:47:1C:D2:81:87:CA:A1:6A:E6, AKID=BC:42:B1:AC:E1:64:5A:31:96:B1:4D:DA:91:1D:5C:CE:CA:10:21:26 failed: x509: certificate signed by unknown authority]" 2025-12-04T11:58:02.191381683Z E1204 11:58:02.191335 1 authentication.go:73] "Unable to authenticate the request" err="[x509: certificate signed by unknown authority, verifying certificate SN=4812284156861786932, SKID=03:5A:81:6F:76:6B:16:1A:1E:F5:A6:47:1C:D2:81:87:CA:A1:6A:E6, AKID=BC:42:B1:AC:E1:64:5A:31:96:B1:4D:DA:91:1D:5C:CE:CA:10:21:26 failed: x509: certificate signed by unknown authority]" 2025-12-04T11:58:26.213282553Z E1204 11:58:26.212001 1 authentication.go:73] "Unable to authenticate the request" err="[x509: certificate signed by unknown authority, verifying certificate SN=4812284156861786932, SKID=03:5A:81:6F:76:6B:16:1A:1E:F5:A6:47:1C:D2:81:87:CA:A1:6A:E6, AKID=BC:42:B1:AC:E1:64:5A:31:96:B1:4D:DA:91:1D:5C:CE:CA:10:21:26 failed: x509: certificate signed by unknown authority]" 2025-12-04T11:58:26.213282553Z E1204 11:58:26.212556 1 authentication.go:73] "Unable to authenticate the request" err="[x509: certificate signed by unknown authority, verifying certificate SN=4812284156861786932, SKID=03:5A:81:6F:76:6B:16:1A:1E:F5:A6:47:1C:D2:81:87:CA:A1:6A:E6, AKID=BC:42:B1:AC:E1:64:5A:31:96:B1:4D:DA:91:1D:5C:CE:CA:10:21:26 failed: x509: certificate signed by unknown authority]" 2025-12-04T11:58:26.213282553Z E1204 11:58:26.212797 1 authentication.go:73] "Unable to authenticate the request" err="[x509: certificate signed by unknown authority, verifying certificate SN=4812284156861786932, SKID=03:5A:81:6F:76:6B:16:1A:1E:F5:A6:47:1C:D2:81:87:CA:A1:6A:E6, AKID=BC:42:B1:AC:E1:64:5A:31:96:B1:4D:DA:91:1D:5C:CE:CA:10:21:26 failed: x509: certificate signed by unknown authority]" 2025-12-04T11:58:26.213282553Z E1204 11:58:26.213069 1 authentication.go:73] "Unable to authenticate the request" err="[x509: certificate signed by unknown authority, verifying certificate SN=4812284156861786932, SKID=03:5A:81:6F:76:6B:16:1A:1E:F5:A6:47:1C:D2:81:87:CA:A1:6A:E6, AKID=BC:42:B1:AC:E1:64:5A:31:96:B1:4D:DA:91:1D:5C:CE:CA:10:21:26 failed: x509: certificate signed by unknown authority]" 2025-12-04T11:58:26.213282553Z E1204 11:58:26.213142 1 authentication.go:73] "Unable to authenticate the request" err="[x509: certificate signed by unknown authority, verifying certificate SN=4812284156861786932, SKID=03:5A:81:6F:76:6B:16:1A:1E:F5:A6:47:1C:D2:81:87:CA:A1:6A:E6, AKID=BC:42:B1:AC:E1:64:5A:31:96:B1:4D:DA:91:1D:5C:CE:CA:10:21:26 failed: x509: certificate signed by unknown authority]" 2025-12-04T11:58:26.213355145Z E1204 11:58:26.213297 1 authentication.go:73] "Unable to authenticate the request" err="[x509: certificate signed by unknown authority, verifying certificate SN=4812284156861786932, SKID=03:5A:81:6F:76:6B:16:1A:1E:F5:A6:47:1C:D2:81:87:CA:A1:6A:E6, AKID=BC:42:B1:AC:E1:64:5A:31:96:B1:4D:DA:91:1D:5C:CE:CA:10:21:26 failed: x509: certificate signed by unknown authority]" 2025-12-04T11:58:26.216440718Z E1204 11:58:26.213520 1 authentication.go:73] "Unable to authenticate the request" err="[x509: certificate signed by unknown authority, verifying certificate SN=4812284156861786932, SKID=03:5A:81:6F:76:6B:16:1A:1E:F5:A6:47:1C:D2:81:87:CA:A1:6A:E6, AKID=BC:42:B1:AC:E1:64:5A:31:96:B1:4D:DA:91:1D:5C:CE:CA:10:21:26 failed: x509: certificate signed by unknown authority]" 2025-12-04T11:58:26.216440718Z E1204 11:58:26.213710 1 authentication.go:73] "Unable to authenticate the request" err="[x509: certificate signed by unknown authority, verifying certificate SN=4812284156861786932, SKID=03:5A:81:6F:76:6B:16:1A:1E:F5:A6:47:1C:D2:81:87:CA:A1:6A:E6, AKID=BC:42:B1:AC:E1:64:5A:31:96:B1:4D:DA:91:1D:5C:CE:CA:10:21:26 failed: x509: certificate signed by unknown authority]" 2025-12-04T11:58:26.216440718Z E1204 11:58:26.213751 1 authentication.go:73] "Unable to authenticate the request" err="[x509: certificate signed by unknown authority, verifying certificate SN=4812284156861786932, SKID=03:5A:81:6F:76:6B:16:1A:1E:F5:A6:47:1C:D2:81:87:CA:A1:6A:E6, AKID=BC:42:B1:AC:E1:64:5A:31:96:B1:4D:DA:91:1D:5C:CE:CA:10:21:26 failed: x509: certificate signed by unknown authority]" 2025-12-04T11:58:26.216440718Z E1204 11:58:26.213917 1 authentication.go:73] "Unable to authenticate the request" err="[x509: certificate signed by unknown authority, verifying certificate SN=4812284156861786932, SKID=03:5A:81:6F:76:6B:16:1A:1E:F5:A6:47:1C:D2:81:87:CA:A1:6A:E6, AKID=BC:42:B1:AC:E1:64:5A:31:96:B1:4D:DA:91:1D:5C:CE:CA:10:21:26 failed: x509: certificate signed by unknown authority]" 2025-12-04T11:58:26.282018190Z E1204 11:58:26.281886 1 authentication.go:73] "Unable to authenticate the request" err="[x509: certificate signed by unknown authority, verifying certificate SN=4812284156861786932, SKID=03:5A:81:6F:76:6B:16:1A:1E:F5:A6:47:1C:D2:81:87:CA:A1:6A:E6, AKID=BC:42:B1:AC:E1:64:5A:31:96:B1:4D:DA:91:1D:5C:CE:CA:10:21:26 failed: x509: certificate signed by unknown authority]" 2025-12-04T11:58:26.289360107Z E1204 11:58:26.289291 1 authentication.go:73] "Unable to authenticate the request" err="[x509: certificate signed by unknown authority, verifying certificate SN=4812284156861786932, SKID=03:5A:81:6F:76:6B:16:1A:1E:F5:A6:47:1C:D2:81:87:CA:A1:6A:E6, AKID=BC:42:B1:AC:E1:64:5A:31:96:B1:4D:DA:91:1D:5C:CE:CA:10:21:26 failed: x509: certificate signed by unknown authority]" 2025-12-04T11:58:26.307205513Z E1204 11:58:26.307128 1 authentication.go:73] "Unable to authenticate the request" err="[x509: certificate signed by unknown authority, verifying certificate SN=4812284156861786932, SKID=03:5A:81:6F:76:6B:16:1A:1E:F5:A6:47:1C:D2:81:87:CA:A1:6A:E6, AKID=BC:42:B1:AC:E1:64:5A:31:96:B1:4D:DA:91:1D:5C:CE:CA:10:21:26 failed: x509: certificate signed by unknown authority]" 2025-12-04T11:58:26.315348161Z E1204 11:58:26.315296 1 authentication.go:73] "Unable to authenticate the request" err="[x509: certificate signed by unknown authority, verifying certificate SN=4812284156861786932, SKID=03:5A:81:6F:76:6B:16:1A:1E:F5:A6:47:1C:D2:81:87:CA:A1:6A:E6, AKID=BC:42:B1:AC:E1:64:5A:31:96:B1:4D:DA:91:1D:5C:CE:CA:10:21:26 failed: x509: certificate signed by unknown authority]" 2025-12-04T11:58:32.160380418Z E1204 11:58:32.160290 1 authentication.go:73] "Unable to authenticate the request" err="[x509: certificate signed by unknown authority, verifying certificate SN=4812284156861786932, SKID=03:5A:81:6F:76:6B:16:1A:1E:F5:A6:47:1C:D2:81:87:CA:A1:6A:E6, AKID=BC:42:B1:AC:E1:64:5A:31:96:B1:4D:DA:91:1D:5C:CE:CA:10:21:26 failed: x509: certificate signed by unknown authority]" 2025-12-04T11:58:32.160380418Z E1204 11:58:32.160340 1 authentication.go:73] "Unable to authenticate the request" err="[x509: certificate signed by unknown authority, verifying certificate SN=4812284156861786932, SKID=03:5A:81:6F:76:6B:16:1A:1E:F5:A6:47:1C:D2:81:87:CA:A1:6A:E6, AKID=BC:42:B1:AC:E1:64:5A:31:96:B1:4D:DA:91:1D:5C:CE:CA:10:21:26 failed: x509: certificate signed by unknown authority]" 2025-12-04T11:58:32.160440690Z E1204 11:58:32.160389 1 authentication.go:73] "Unable to authenticate the request" err="[x509: certificate signed by unknown authority, verifying certificate SN=4812284156861786932, SKID=03:5A:81:6F:76:6B:16:1A:1E:F5:A6:47:1C:D2:81:87:CA:A1:6A:E6, AKID=BC:42:B1:AC:E1:64:5A:31:96:B1:4D:DA:91:1D:5C:CE:CA:10:21:26 failed: x509: certificate signed by unknown authority]" 2025-12-04T11:58:32.160553853Z E1204 11:58:32.160290 1 authentication.go:73] "Unable to authenticate the request" err="[x509: certificate signed by unknown authority, verifying certificate SN=4812284156861786932, SKID=03:5A:81:6F:76:6B:16:1A:1E:F5:A6:47:1C:D2:81:87:CA:A1:6A:E6, AKID=BC:42:B1:AC:E1:64:5A:31:96:B1:4D:DA:91:1D:5C:CE:CA:10:21:26 failed: x509: certificate signed by unknown authority]" 2025-12-04T11:58:32.160646766Z E1204 11:58:32.160618 1 authentication.go:73] "Unable to authenticate the request" err="[x509: certificate signed by unknown authority, verifying certificate SN=4812284156861786932, SKID=03:5A:81:6F:76:6B:16:1A:1E:F5:A6:47:1C:D2:81:87:CA:A1:6A:E6, AKID=BC:42:B1:AC:E1:64:5A:31:96:B1:4D:DA:91:1D:5C:CE:CA:10:21:26 failed: x509: certificate signed by unknown authority]" 2025-12-04T11:58:32.194650620Z E1204 11:58:32.194561 1 authentication.go:73] "Unable to authenticate the request" err="[x509: certificate signed by unknown authority, verifying certificate SN=4812284156861786932, SKID=03:5A:81:6F:76:6B:16:1A:1E:F5:A6:47:1C:D2:81:87:CA:A1:6A:E6, AKID=BC:42:B1:AC:E1:64:5A:31:96:B1:4D:DA:91:1D:5C:CE:CA:10:21:26 failed: x509: certificate signed by unknown authority]" 2025-12-04T11:58:32.195112185Z E1204 11:58:32.195029 1 authentication.go:73] "Unable to authenticate the request" err="[x509: certificate signed by unknown authority, verifying certificate SN=4812284156861786932, SKID=03:5A:81:6F:76:6B:16:1A:1E:F5:A6:47:1C:D2:81:87:CA:A1:6A:E6, AKID=BC:42:B1:AC:E1:64:5A:31:96:B1:4D:DA:91:1D:5C:CE:CA:10:21:26 failed: x509: certificate signed by unknown authority]" 2025-12-04T11:58:32.195570700Z E1204 11:58:32.195048 1 authentication.go:73] "Unable to authenticate the request" err="[x509: certificate signed by unknown authority, verifying certificate SN=4812284156861786932, SKID=03:5A:81:6F:76:6B:16:1A:1E:F5:A6:47:1C:D2:81:87:CA:A1:6A:E6, AKID=BC:42:B1:AC:E1:64:5A:31:96:B1:4D:DA:91:1D:5C:CE:CA:10:21:26 failed: x509: certificate signed by unknown authority]" 2025-12-04T11:58:32.195570700Z E1204 11:58:32.195275 1 authentication.go:73] "Unable to authenticate the request" err="[x509: certificate signed by unknown authority, verifying certificate SN=4812284156861786932, SKID=03:5A:81:6F:76:6B:16:1A:1E:F5:A6:47:1C:D2:81:87:CA:A1:6A:E6, AKID=BC:42:B1:AC:E1:64:5A:31:96:B1:4D:DA:91:1D:5C:CE:CA:10:21:26 failed: x509: certificate signed by unknown authority]" 2025-12-04T11:58:32.195570700Z E1204 11:58:32.195048 1 authentication.go:73] "Unable to authenticate the request" err="[x509: certificate signed by unknown authority, verifying certificate SN=4812284156861786932, SKID=03:5A:81:6F:76:6B:16:1A:1E:F5:A6:47:1C:D2:81:87:CA:A1:6A:E6, AKID=BC:42:B1:AC:E1:64:5A:31:96:B1:4D:DA:91:1D:5C:CE:CA:10:21:26 failed: x509: certificate signed by unknown authority]" 2025-12-04T11:58:35.728683968Z I1204 11:58:35.728590 1 policy_source.go:224] refreshing policies 2025-12-04T11:58:56.228172185Z E1204 11:58:56.228091 1 authentication.go:73] "Unable to authenticate the request" err="[x509: certificate signed by unknown authority, verifying certificate SN=4812284156861786932, SKID=03:5A:81:6F:76:6B:16:1A:1E:F5:A6:47:1C:D2:81:87:CA:A1:6A:E6, AKID=BC:42:B1:AC:E1:64:5A:31:96:B1:4D:DA:91:1D:5C:CE:CA:10:21:26 failed: x509: certificate signed by unknown authority]" 2025-12-04T11:58:56.228617309Z E1204 11:58:56.228571 1 authentication.go:73] "Unable to authenticate the request" err="[x509: certificate signed by unknown authority, verifying certificate SN=4812284156861786932, SKID=03:5A:81:6F:76:6B:16:1A:1E:F5:A6:47:1C:D2:81:87:CA:A1:6A:E6, AKID=BC:42:B1:AC:E1:64:5A:31:96:B1:4D:DA:91:1D:5C:CE:CA:10:21:26 failed: x509: certificate signed by unknown authority]" 2025-12-04T11:58:56.228978190Z E1204 11:58:56.228841 1 authentication.go:73] "Unable to authenticate the request" err="[x509: certificate signed by unknown authority, verifying certificate SN=4812284156861786932, SKID=03:5A:81:6F:76:6B:16:1A:1E:F5:A6:47:1C:D2:81:87:CA:A1:6A:E6, AKID=BC:42:B1:AC:E1:64:5A:31:96:B1:4D:DA:91:1D:5C:CE:CA:10:21:26 failed: x509: certificate signed by unknown authority]" 2025-12-04T11:58:56.229122535Z E1204 11:58:56.229076 1 authentication.go:73] "Unable to authenticate the request" err="[x509: certificate signed by unknown authority, verifying certificate SN=4812284156861786932, SKID=03:5A:81:6F:76:6B:16:1A:1E:F5:A6:47:1C:D2:81:87:CA:A1:6A:E6, AKID=BC:42:B1:AC:E1:64:5A:31:96:B1:4D:DA:91:1D:5C:CE:CA:10:21:26 failed: x509: certificate signed by unknown authority]" 2025-12-04T11:58:56.229345652Z E1204 11:58:56.229303 1 authentication.go:73] "Unable to authenticate the request" err="[x509: certificate signed by unknown authority, verifying certificate SN=4812284156861786932, SKID=03:5A:81:6F:76:6B:16:1A:1E:F5:A6:47:1C:D2:81:87:CA:A1:6A:E6, AKID=BC:42:B1:AC:E1:64:5A:31:96:B1:4D:DA:91:1D:5C:CE:CA:10:21:26 failed: x509: certificate signed by unknown authority]" 2025-12-04T11:58:56.229561189Z E1204 11:58:56.229515 1 authentication.go:73] "Unable to authenticate the request" err="[x509: certificate signed by unknown authority, verifying certificate SN=4812284156861786932, SKID=03:5A:81:6F:76:6B:16:1A:1E:F5:A6:47:1C:D2:81:87:CA:A1:6A:E6, AKID=BC:42:B1:AC:E1:64:5A:31:96:B1:4D:DA:91:1D:5C:CE:CA:10:21:26 failed: x509: certificate signed by unknown authority]" 2025-12-04T11:58:56.229733995Z E1204 11:58:56.229692 1 authentication.go:73] "Unable to authenticate the request" err="[x509: certificate signed by unknown authority, verifying certificate SN=4812284156861786932, SKID=03:5A:81:6F:76:6B:16:1A:1E:F5:A6:47:1C:D2:81:87:CA:A1:6A:E6, AKID=BC:42:B1:AC:E1:64:5A:31:96:B1:4D:DA:91:1D:5C:CE:CA:10:21:26 failed: x509: certificate signed by unknown authority]" 2025-12-04T11:58:56.229974382Z E1204 11:58:56.229924 1 authentication.go:73] "Unable to authenticate the request" err="[x509: certificate signed by unknown authority, verifying certificate SN=4812284156861786932, SKID=03:5A:81:6F:76:6B:16:1A:1E:F5:A6:47:1C:D2:81:87:CA:A1:6A:E6, AKID=BC:42:B1:AC:E1:64:5A:31:96:B1:4D:DA:91:1D:5C:CE:CA:10:21:26 failed: x509: certificate signed by unknown authority]" 2025-12-04T11:58:56.230202499Z E1204 11:58:56.230169 1 authentication.go:73] "Unable to authenticate the request" err="[x509: certificate signed by unknown authority, verifying certificate SN=4812284156861786932, SKID=03:5A:81:6F:76:6B:16:1A:1E:F5:A6:47:1C:D2:81:87:CA:A1:6A:E6, AKID=BC:42:B1:AC:E1:64:5A:31:96:B1:4D:DA:91:1D:5C:CE:CA:10:21:26 failed: x509: certificate signed by unknown authority]" 2025-12-04T11:58:56.230473058Z E1204 11:58:56.230432 1 authentication.go:73] "Unable to authenticate the request" err="[x509: certificate signed by unknown authority, verifying certificate SN=4812284156861786932, SKID=03:5A:81:6F:76:6B:16:1A:1E:F5:A6:47:1C:D2:81:87:CA:A1:6A:E6, AKID=BC:42:B1:AC:E1:64:5A:31:96:B1:4D:DA:91:1D:5C:CE:CA:10:21:26 failed: x509: certificate signed by unknown authority]" 2025-12-04T11:59:02.168861310Z E1204 11:59:02.168390 1 authentication.go:73] "Unable to authenticate the request" err="[x509: certificate signed by unknown authority, verifying certificate SN=4812284156861786932, SKID=03:5A:81:6F:76:6B:16:1A:1E:F5:A6:47:1C:D2:81:87:CA:A1:6A:E6, AKID=BC:42:B1:AC:E1:64:5A:31:96:B1:4D:DA:91:1D:5C:CE:CA:10:21:26 failed: x509: certificate signed by unknown authority]" 2025-12-04T11:59:02.168861310Z E1204 11:59:02.168518 1 authentication.go:73] "Unable to authenticate the request" err="[x509: certificate signed by unknown authority, verifying certificate SN=4812284156861786932, SKID=03:5A:81:6F:76:6B:16:1A:1E:F5:A6:47:1C:D2:81:87:CA:A1:6A:E6, AKID=BC:42:B1:AC:E1:64:5A:31:96:B1:4D:DA:91:1D:5C:CE:CA:10:21:26 failed: x509: certificate signed by unknown authority]" 2025-12-04T11:59:02.168861310Z E1204 11:59:02.168728 1 authentication.go:73] "Unable to authenticate the request" err="[x509: certificate signed by unknown authority, verifying certificate SN=4812284156861786932, SKID=03:5A:81:6F:76:6B:16:1A:1E:F5:A6:47:1C:D2:81:87:CA:A1:6A:E6, AKID=BC:42:B1:AC:E1:64:5A:31:96:B1:4D:DA:91:1D:5C:CE:CA:10:21:26 failed: x509: certificate signed by unknown authority]" 2025-12-04T11:59:02.172845807Z E1204 11:59:02.168958 1 authentication.go:73] "Unable to authenticate the request" err="[x509: certificate signed by unknown authority, verifying certificate SN=4812284156861786932, SKID=03:5A:81:6F:76:6B:16:1A:1E:F5:A6:47:1C:D2:81:87:CA:A1:6A:E6, AKID=BC:42:B1:AC:E1:64:5A:31:96:B1:4D:DA:91:1D:5C:CE:CA:10:21:26 failed: x509: certificate signed by unknown authority]" 2025-12-04T11:59:02.172845807Z E1204 11:59:02.169080 1 authentication.go:73] "Unable to authenticate the request" err="[x509: certificate signed by unknown authority, verifying certificate SN=4812284156861786932, SKID=03:5A:81:6F:76:6B:16:1A:1E:F5:A6:47:1C:D2:81:87:CA:A1:6A:E6, AKID=BC:42:B1:AC:E1:64:5A:31:96:B1:4D:DA:91:1D:5C:CE:CA:10:21:26 failed: x509: certificate signed by unknown authority]" 2025-12-04T11:59:02.204426937Z E1204 11:59:02.204321 1 authentication.go:73] "Unable to authenticate the request" err="[x509: certificate signed by unknown authority, verifying certificate SN=4812284156861786932, SKID=03:5A:81:6F:76:6B:16:1A:1E:F5:A6:47:1C:D2:81:87:CA:A1:6A:E6, AKID=BC:42:B1:AC:E1:64:5A:31:96:B1:4D:DA:91:1D:5C:CE:CA:10:21:26 failed: x509: certificate signed by unknown authority]" 2025-12-04T11:59:02.204426937Z E1204 11:59:02.204402 1 authentication.go:73] "Unable to authenticate the request" err="[x509: certificate signed by unknown authority, verifying certificate SN=4812284156861786932, SKID=03:5A:81:6F:76:6B:16:1A:1E:F5:A6:47:1C:D2:81:87:CA:A1:6A:E6, AKID=BC:42:B1:AC:E1:64:5A:31:96:B1:4D:DA:91:1D:5C:CE:CA:10:21:26 failed: x509: certificate signed by unknown authority]" 2025-12-04T11:59:02.204426937Z E1204 11:59:02.204405 1 authentication.go:73] "Unable to authenticate the request" err="[x509: certificate signed by unknown authority, verifying certificate SN=4812284156861786932, SKID=03:5A:81:6F:76:6B:16:1A:1E:F5:A6:47:1C:D2:81:87:CA:A1:6A:E6, AKID=BC:42:B1:AC:E1:64:5A:31:96:B1:4D:DA:91:1D:5C:CE:CA:10:21:26 failed: x509: certificate signed by unknown authority]" 2025-12-04T11:59:02.204536671Z E1204 11:59:02.204321 1 authentication.go:73] "Unable to authenticate the request" err="[x509: certificate signed by unknown authority, verifying certificate SN=4812284156861786932, SKID=03:5A:81:6F:76:6B:16:1A:1E:F5:A6:47:1C:D2:81:87:CA:A1:6A:E6, AKID=BC:42:B1:AC:E1:64:5A:31:96:B1:4D:DA:91:1D:5C:CE:CA:10:21:26 failed: x509: certificate signed by unknown authority]" 2025-12-04T11:59:02.204651324Z E1204 11:59:02.204321 1 authentication.go:73] "Unable to authenticate the request" err="[x509: certificate signed by unknown authority, verifying certificate SN=4812284156861786932, SKID=03:5A:81:6F:76:6B:16:1A:1E:F5:A6:47:1C:D2:81:87:CA:A1:6A:E6, AKID=BC:42:B1:AC:E1:64:5A:31:96:B1:4D:DA:91:1D:5C:CE:CA:10:21:26 failed: x509: certificate signed by unknown authority]" 2025-12-04T11:59:21.138976932Z E1204 11:59:21.138887 1 webhook.go:269] Failed to make webhook authorizer request: Post "https://172.30.0.1:443/apis/authorization.k8s.io/v1/subjectaccessreviews?timeout=10s": dial tcp 172.30.0.1:443: connect: connection refused 2025-12-04T11:59:21.139051504Z E1204 11:59:21.139010 1 errors.go:77] "Unhandled Error" err="Post \"https://172.30.0.1:443/apis/authorization.k8s.io/v1/subjectaccessreviews?timeout=10s\": dial tcp 172.30.0.1:443: connect: connection refused" 2025-12-04T11:59:21.178416951Z E1204 11:59:21.178336 1 webhook.go:269] Failed to make webhook authorizer request: Post "https://172.30.0.1:443/apis/authorization.k8s.io/v1/subjectaccessreviews?timeout=10s": dial tcp 172.30.0.1:443: connect: connection refused 2025-12-04T11:59:21.178416951Z E1204 11:59:21.178402 1 errors.go:77] "Unhandled Error" err="Post \"https://172.30.0.1:443/apis/authorization.k8s.io/v1/subjectaccessreviews?timeout=10s\": dial tcp 172.30.0.1:443: connect: connection refused" 2025-12-04T11:59:22.007938783Z E1204 11:59:22.007729 1 webhook.go:269] Failed to make webhook authorizer request: Post "https://172.30.0.1:443/apis/authorization.k8s.io/v1/subjectaccessreviews?timeout=10s": dial tcp 172.30.0.1:443: connect: connection refused 2025-12-04T11:59:22.007938783Z E1204 11:59:22.007835 1 errors.go:77] "Unhandled Error" err="Post \"https://172.30.0.1:443/apis/authorization.k8s.io/v1/subjectaccessreviews?timeout=10s\": dial tcp 172.30.0.1:443: connect: connection refused" 2025-12-04T11:59:22.016905527Z E1204 11:59:22.016736 1 webhook.go:269] Failed to make webhook authorizer request: Post "https://172.30.0.1:443/apis/authorization.k8s.io/v1/subjectaccessreviews?timeout=10s": dial tcp 172.30.0.1:443: connect: connection refused 2025-12-04T11:59:22.016905527Z E1204 11:59:22.016797 1 errors.go:77] "Unhandled Error" err="Post \"https://172.30.0.1:443/apis/authorization.k8s.io/v1/subjectaccessreviews?timeout=10s\": dial tcp 172.30.0.1:443: connect: connection refused" 2025-12-04T11:59:22.028512065Z E1204 11:59:22.026344 1 webhook.go:269] Failed to make webhook authorizer request: Post "https://172.30.0.1:443/apis/authorization.k8s.io/v1/subjectaccessreviews?timeout=10s": dial tcp 172.30.0.1:443: connect: connection refused 2025-12-04T11:59:22.028512065Z E1204 11:59:22.026411 1 errors.go:77] "Unhandled Error" err="Post \"https://172.30.0.1:443/apis/authorization.k8s.io/v1/subjectaccessreviews?timeout=10s\": dial tcp 172.30.0.1:443: connect: connection refused" 2025-12-04T11:59:22.035264088Z E1204 11:59:22.035161 1 webhook.go:269] Failed to make webhook authorizer request: Post "https://172.30.0.1:443/apis/authorization.k8s.io/v1/subjectaccessreviews?timeout=10s": dial tcp 172.30.0.1:443: connect: connection refused 2025-12-04T11:59:22.035264088Z E1204 11:59:22.035237 1 errors.go:77] "Unhandled Error" err="Post \"https://172.30.0.1:443/apis/authorization.k8s.io/v1/subjectaccessreviews?timeout=10s\": dial tcp 172.30.0.1:443: connect: connection refused" 2025-12-04T11:59:46.245326967Z I1204 11:59:46.245257 1 reflector.go:368] Caches populated for *v1.Namespace from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243 2025-12-04T11:59:52.703436154Z I1204 11:59:52.703305 1 reflector.go:368] Caches populated for *v1.ValidatingAdmissionPolicyBinding from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243 2025-12-04T11:59:52.780004738Z I1204 11:59:52.779841 1 policy_source.go:224] refreshing policies 2025-12-04T11:59:55.917753564Z I1204 11:59:55.917652 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243 2025-12-04T11:59:56.900015713Z I1204 11:59:56.899907 1 reflector.go:368] Caches populated for *v1.MutatingWebhookConfiguration from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243 2025-12-04T11:59:57.151374933Z I1204 11:59:57.151276 1 reflector.go:368] Caches populated for *v1.ValidatingAdmissionPolicy from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243 2025-12-04T11:59:57.805704938Z I1204 11:59:57.805629 1 policy_source.go:224] refreshing policies 2025-12-04T11:59:59.215588009Z I1204 11:59:59.215469 1 reflector.go:368] Caches populated for *v1.ValidatingWebhookConfiguration from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243 2025-12-04T12:00:00.554670069Z I1204 12:00:00.554603 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243 2025-12-04T12:00:02.216954475Z I1204 12:00:02.216858 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243 2025-12-04T12:00:02.217628737Z I1204 12:00:02.217562 1 tlsconfig.go:181] "Loaded client CA" index=0 certName="client-ca::kube-system::extension-apiserver-authentication::client-ca-file,client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file" certDetail="\"admin-kubeconfig-signer\" [] issuer=\"\" (2025-12-04 11:20:50 +0000 UTC to 2035-12-02 11:20:50 +0000 UTC (now=2025-12-04 12:00:02.217513073 +0000 UTC))" 2025-12-04T12:00:02.217628737Z I1204 12:00:02.217616 1 tlsconfig.go:181] "Loaded client CA" index=1 certName="client-ca::kube-system::extension-apiserver-authentication::client-ca-file,client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file" certDetail="\"kubelet-signer\" [] issuer=\"\" (2025-12-04 11:26:53 +0000 UTC to 2025-12-05 11:26:53 +0000 UTC (now=2025-12-04 12:00:02.217601856 +0000 UTC))" 2025-12-04T12:00:02.217643227Z I1204 12:00:02.217630 1 tlsconfig.go:181] "Loaded client CA" index=2 certName="client-ca::kube-system::extension-apiserver-authentication::client-ca-file,client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file" certDetail="\"kube-control-plane-signer\" [] issuer=\"\" (2025-12-04 11:26:54 +0000 UTC to 2026-12-04 11:26:54 +0000 UTC (now=2025-12-04 12:00:02.217621577 +0000 UTC))" 2025-12-04T12:00:02.217651328Z I1204 12:00:02.217642 1 tlsconfig.go:181] "Loaded client CA" index=3 certName="client-ca::kube-system::extension-apiserver-authentication::client-ca-file,client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file" certDetail="\"kube-apiserver-to-kubelet-signer\" [] issuer=\"\" (2025-12-04 11:26:54 +0000 UTC to 2026-12-04 11:26:54 +0000 UTC (now=2025-12-04 12:00:02.217634537 +0000 UTC))" 2025-12-04T12:00:02.217682759Z I1204 12:00:02.217656 1 tlsconfig.go:181] "Loaded client CA" index=4 certName="client-ca::kube-system::extension-apiserver-authentication::client-ca-file,client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file" certDetail="\"kubelet-bootstrap-kubeconfig-signer\" [] issuer=\"\" (2025-12-04 11:26:52 +0000 UTC to 2035-12-02 11:26:52 +0000 UTC (now=2025-12-04 12:00:02.217646987 +0000 UTC))" 2025-12-04T12:00:02.217682759Z I1204 12:00:02.217673 1 tlsconfig.go:181] "Loaded client CA" index=5 certName="client-ca::kube-system::extension-apiserver-authentication::client-ca-file,client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file" certDetail="\"kube-csr-signer_@1764848267\" [] issuer=\"kubelet-signer\" (2025-12-04 11:37:46 +0000 UTC to 2025-12-05 11:26:53 +0000 UTC (now=2025-12-04 12:00:02.217666058 +0000 UTC))" 2025-12-04T12:00:02.217726000Z I1204 12:00:02.217695 1 tlsconfig.go:181] "Loaded client CA" index=6 certName="client-ca::kube-system::extension-apiserver-authentication::client-ca-file,client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file" certDetail="\"openshift-kube-apiserver-operator_node-system-admin-signer@1764848255\" [] issuer=\"\" (2025-12-04 11:37:34 +0000 UTC to 2026-12-04 11:37:35 +0000 UTC (now=2025-12-04 12:00:02.217677568 +0000 UTC))" 2025-12-04T12:00:02.217726000Z I1204 12:00:02.217721 1 tlsconfig.go:181] "Loaded client CA" index=7 certName="client-ca::kube-system::extension-apiserver-authentication::client-ca-file,client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file" certDetail="\"aggregator-signer\" [] issuer=\"\" (2025-12-04 11:26:52 +0000 UTC to 2025-12-05 11:26:52 +0000 UTC (now=2025-12-04 12:00:02.217710809 +0000 UTC))" 2025-12-04T12:00:02.217936477Z I1204 12:00:02.217910 1 tlsconfig.go:203] "Loaded serving cert" certName="serving-cert::/var/run/secrets/serving-cert/tls.crt::/var/run/secrets/serving-cert/tls.key" certDetail="\"api.openshift-oauth-apiserver.svc\" [serving] validServingFor=[api.openshift-oauth-apiserver.svc,api.openshift-oauth-apiserver.svc.cluster.local] issuer=\"openshift-service-serving-signer@1764848266\" (2025-12-04 11:38:07 +0000 UTC to 2027-12-04 11:38:08 +0000 UTC (now=2025-12-04 12:00:02.217893075 +0000 UTC))" 2025-12-04T12:00:02.218057101Z I1204 12:00:02.218023 1 named_certificates.go:53] "Loaded SNI cert" index=0 certName="self-signed loopback" certDetail="\"apiserver-loopback-client@1764848314\" [serving] validServingFor=[apiserver-loopback-client] issuer=\"apiserver-loopback-client-ca@1764848314\" (2025-12-04 10:38:34 +0000 UTC to 2026-12-04 10:38:34 +0000 UTC (now=2025-12-04 12:00:02.218005879 +0000 UTC))" authentication-operator_current.log0000640000000000000000000011432115114274105037713 0ustar0000000000000000config/pod/openshift-authentication-operator/logs/authentication-operator-6c968fdfdf-nrrfw2025-12-04T11:59:57.913275140Z E1204 11:59:57.913185 1 base_controller.go:279] "Unhandled Error" err="OAuthServerRouteEndpointAccessibleController reconciliation failed: route \"openshift-authentication/oauth-openshift\": status does not have a valid host address" 2025-12-04T11:59:57.918450564Z E1204 11:59:57.918369 1 base_controller.go:279] "Unhandled Error" err="OAuthServerServiceEndpointAccessibleController reconciliation failed: Get \"https://172.30.202.245:443/healthz\": dial tcp 172.30.202.245:443: connect: connection refused" 2025-12-04T11:59:58.031379144Z I1204 11:59:58.030430 1 status_controller.go:225] clusteroperator/authentication diff {"status":{"conditions":[{"lastTransitionTime":"2025-12-04T11:56:30Z","message":"IngressStateEndpointsDegraded: No subsets found for the endpoints of oauth-server\nOAuthServerDeploymentDegraded: waiting for the oauth-openshift route to contain an admitted ingress: no admitted ingress for route oauth-openshift in namespace openshift-authentication\nOAuthServerDeploymentDegraded: \nOAuthServerRouteEndpointAccessibleControllerDegraded: route \"openshift-authentication/oauth-openshift\": status does not have a valid host address\nOAuthServerServiceEndpointAccessibleControllerDegraded: Get \"https://172.30.202.245:443/healthz\": dial tcp 172.30.202.245:443: connect: connection refused\nOAuthServerServiceEndpointsEndpointAccessibleControllerDegraded: oauth service endpoints are not ready\nOAuthServerWorkloadDegraded: waiting for the oauth-openshift route to contain an admitted ingress: no admitted ingress for route oauth-openshift in namespace openshift-authentication\nOAuthServerWorkloadDegraded: \nWellKnownReadyControllerDegraded: failed to get oauth metadata from openshift-config-managed/oauth-openshift ConfigMap: configmap \"oauth-openshift\" not found (check authentication operator, it is supposed to create this)","reason":"IngressStateEndpoints_MissingSubsets::OAuthServerDeployment_PreconditionNotFulfilled::OAuthServerRouteEndpointAccessibleController_SyncError::OAuthServerServiceEndpointAccessibleController_SyncError::OAuthServerServiceEndpointsEndpointAccessibleController_SyncError::OAuthServerWorkload_PreconditionNotFulfilled::WellKnownReadyController_SyncError","status":"True","type":"Degraded"},{"lastTransitionTime":"2025-12-04T11:59:50Z","reason":"WellKnownReadyController","status":"True","type":"Progressing"},{"lastTransitionTime":"2025-12-04T11:37:45Z","message":"OAuthServerRouteEndpointAccessibleControllerAvailable: failed to retrieve route from cache: route.route.openshift.io \"oauth-openshift\" not found\nOAuthServerServiceEndpointAccessibleControllerAvailable: Get \"https://172.30.202.245:443/healthz\": dial tcp 172.30.202.245:443: connect: connection refused\nOAuthServerServiceEndpointsEndpointAccessibleControllerAvailable: endpoints \"oauth-openshift\" not found\nReadyIngressNodesAvailable: Authentication requires functional ingress which requires at least one schedulable and ready node. Got 0 worker nodes, 1 master nodes, 0 custom target nodes (none are schedulable or ready for ingress pods).\nWellKnownAvailable: The well-known endpoint is not yet available: failed to get oauth metadata from openshift-config-managed/oauth-openshift ConfigMap: configmap \"oauth-openshift\" not found (check authentication operator, it is supposed to create this)","reason":"OAuthServerDeployment_PreconditionNotFulfilled::OAuthServerRouteEndpointAccessibleController_ResourceNotFound::OAuthServerServiceEndpointAccessibleController_EndpointUnavailable::OAuthServerServiceEndpointsEndpointAccessibleController_ResourceNotFound::ReadyIngressNodes_NoReadyIngressNodes::WellKnown_NotReady","status":"False","type":"Available"},{"lastTransitionTime":"2025-12-04T11:37:45Z","message":"All is well","reason":"AsExpected","status":"True","type":"Upgradeable"},{"lastTransitionTime":"2025-12-04T11:37:45Z","reason":"NoData","status":"Unknown","type":"EvaluationConditionsDetected"}]}} 2025-12-04T11:59:58.031379144Z I1204 11:59:58.030700 1 event.go:377] Event(v1.ObjectReference{Kind:"Deployment", Namespace:"openshift-authentication-operator", Name:"authentication-operator", UID:"", APIVersion:"", ResourceVersion:"", FieldPath:""}): type: 'Normal' reason: 'OperatorStatusChanged' Status for clusteroperator/authentication changed: Degraded message changed from "IngressStateEndpointsDegraded: No subsets found for the endpoints of oauth-server\nOAuthServerDeploymentDegraded: waiting for the oauth-openshift route to contain an admitted ingress: no admitted ingress for route oauth-openshift in namespace openshift-authentication\nOAuthServerDeploymentDegraded: \nOAuthServerRouteEndpointAccessibleControllerDegraded: ingress.config/cluster does not yet have status for the \"openshift-authentication/oauth-openshift\" route\nOAuthServerServiceEndpointAccessibleControllerDegraded: Get \"https://172.30.202.245:443/healthz\": dial tcp 172.30.202.245:443: connect: connection refused\nOAuthServerServiceEndpointsEndpointAccessibleControllerDegraded: oauth service endpoints are not ready\nOAuthServerWorkloadDegraded: waiting for the oauth-openshift route to contain an admitted ingress: no admitted ingress for route oauth-openshift in namespace openshift-authentication\nOAuthServerWorkloadDegraded: \nWellKnownReadyControllerDegraded: failed to get oauth metadata from openshift-config-managed/oauth-openshift ConfigMap: configmap \"oauth-openshift\" not found (check authentication operator, it is supposed to create this)" to "IngressStateEndpointsDegraded: No subsets found for the endpoints of oauth-server\nOAuthServerDeploymentDegraded: waiting for the oauth-openshift route to contain an admitted ingress: no admitted ingress for route oauth-openshift in namespace openshift-authentication\nOAuthServerDeploymentDegraded: \nOAuthServerRouteEndpointAccessibleControllerDegraded: route \"openshift-authentication/oauth-openshift\": status does not have a valid host address\nOAuthServerServiceEndpointAccessibleControllerDegraded: Get \"https://172.30.202.245:443/healthz\": dial tcp 172.30.202.245:443: connect: connection refused\nOAuthServerServiceEndpointsEndpointAccessibleControllerDegraded: oauth service endpoints are not ready\nOAuthServerWorkloadDegraded: waiting for the oauth-openshift route to contain an admitted ingress: no admitted ingress for route oauth-openshift in namespace openshift-authentication\nOAuthServerWorkloadDegraded: \nWellKnownReadyControllerDegraded: failed to get oauth metadata from openshift-config-managed/oauth-openshift ConfigMap: configmap \"oauth-openshift\" not found (check authentication operator, it is supposed to create this)" 2025-12-04T11:59:58.045881763Z E1204 11:59:58.045720 1 base_controller.go:279] "Unhandled Error" err="StatusSyncer_authentication reconciliation failed: Operation cannot be fulfilled on clusteroperators.config.openshift.io \"authentication\": the object has been modified; please apply your changes to the latest version and try again" 2025-12-04T11:59:58.046179102Z I1204 11:59:58.046099 1 reflector.go:368] Caches populated for *v1.Namespace from k8s.io/client-go/informers/factory.go:160 2025-12-04T11:59:58.059961999Z E1204 11:59:58.059848 1 base_controller.go:279] "Unhandled Error" err="OAuthServer-WorkloadWorkloadController reconciliation failed: waiting for the oauth-openshift route to contain an admitted ingress: no admitted ingress for route oauth-openshift in namespace openshift-authentication" 2025-12-04T11:59:58.073455057Z I1204 11:59:58.073365 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/apiserver/pkg/authentication/request/headerrequest/requestheader_controller.go:175 2025-12-04T11:59:58.443161715Z I1204 11:59:58.442955 1 reflector.go:368] Caches populated for *v1.Pod from k8s.io/client-go/informers/factory.go:160 2025-12-04T11:59:58.648123767Z I1204 11:59:58.647339 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go/informers/factory.go:160 2025-12-04T11:59:58.732925053Z I1204 11:59:58.732781 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/apiserver/pkg/server/dynamiccertificates/configmap_cafile_content.go:209 2025-12-04T11:59:58.733578964Z I1204 11:59:58.733500 1 tlsconfig.go:181] "Loaded client CA" index=0 certName="client-ca::kube-system::extension-apiserver-authentication::client-ca-file,client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file" certDetail="\"admin-kubeconfig-signer\" [] issuer=\"\" (2025-12-04 11:20:50 +0000 UTC to 2035-12-02 11:20:50 +0000 UTC (now=2025-12-04 11:59:58.733450609 +0000 UTC))" 2025-12-04T11:59:58.733578964Z I1204 11:59:58.733563 1 tlsconfig.go:181] "Loaded client CA" index=1 certName="client-ca::kube-system::extension-apiserver-authentication::client-ca-file,client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file" certDetail="\"kubelet-signer\" [] issuer=\"\" (2025-12-04 11:26:53 +0000 UTC to 2025-12-05 11:26:53 +0000 UTC (now=2025-12-04 11:59:58.733550563 +0000 UTC))" 2025-12-04T11:59:58.733603884Z I1204 11:59:58.733582 1 tlsconfig.go:181] "Loaded client CA" index=2 certName="client-ca::kube-system::extension-apiserver-authentication::client-ca-file,client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file" certDetail="\"kube-control-plane-signer\" [] issuer=\"\" (2025-12-04 11:26:54 +0000 UTC to 2026-12-04 11:26:54 +0000 UTC (now=2025-12-04 11:59:58.733569433 +0000 UTC))" 2025-12-04T11:59:58.733603884Z I1204 11:59:58.733596 1 tlsconfig.go:181] "Loaded client CA" index=3 certName="client-ca::kube-system::extension-apiserver-authentication::client-ca-file,client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file" certDetail="\"kube-apiserver-to-kubelet-signer\" [] issuer=\"\" (2025-12-04 11:26:54 +0000 UTC to 2026-12-04 11:26:54 +0000 UTC (now=2025-12-04 11:59:58.733587574 +0000 UTC))" 2025-12-04T11:59:58.733648466Z I1204 11:59:58.733610 1 tlsconfig.go:181] "Loaded client CA" index=4 certName="client-ca::kube-system::extension-apiserver-authentication::client-ca-file,client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file" certDetail="\"kubelet-bootstrap-kubeconfig-signer\" [] issuer=\"\" (2025-12-04 11:26:52 +0000 UTC to 2035-12-02 11:26:52 +0000 UTC (now=2025-12-04 11:59:58.733601554 +0000 UTC))" 2025-12-04T11:59:58.733648466Z I1204 11:59:58.733631 1 tlsconfig.go:181] "Loaded client CA" index=5 certName="client-ca::kube-system::extension-apiserver-authentication::client-ca-file,client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file" certDetail="\"kube-csr-signer_@1764848267\" [] issuer=\"kubelet-signer\" (2025-12-04 11:37:46 +0000 UTC to 2025-12-05 11:26:53 +0000 UTC (now=2025-12-04 11:59:58.733621595 +0000 UTC))" 2025-12-04T11:59:58.733681417Z I1204 11:59:58.733655 1 tlsconfig.go:181] "Loaded client CA" index=6 certName="client-ca::kube-system::extension-apiserver-authentication::client-ca-file,client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file" certDetail="\"openshift-kube-apiserver-operator_node-system-admin-signer@1764848255\" [] issuer=\"\" (2025-12-04 11:37:34 +0000 UTC to 2026-12-04 11:37:35 +0000 UTC (now=2025-12-04 11:59:58.733643596 +0000 UTC))" 2025-12-04T11:59:58.733691817Z I1204 11:59:58.733677 1 tlsconfig.go:181] "Loaded client CA" index=7 certName="client-ca::kube-system::extension-apiserver-authentication::client-ca-file,client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file" certDetail="\"aggregator-signer\" [] issuer=\"\" (2025-12-04 11:26:52 +0000 UTC to 2025-12-05 11:26:52 +0000 UTC (now=2025-12-04 11:59:58.733667776 +0000 UTC))" 2025-12-04T11:59:58.733917454Z I1204 11:59:58.733880 1 tlsconfig.go:203] "Loaded serving cert" certName="serving-cert::/var/run/secrets/serving-cert/tls.crt::/var/run/secrets/serving-cert/tls.key" certDetail="\"metrics.openshift-authentication-operator.svc\" [serving] validServingFor=[metrics.openshift-authentication-operator.svc,metrics.openshift-authentication-operator.svc.cluster.local] issuer=\"openshift-service-serving-signer@1764848266\" (2025-12-04 11:37:52 +0000 UTC to 2027-12-04 11:37:53 +0000 UTC (now=2025-12-04 11:59:58.733861143 +0000 UTC))" 2025-12-04T11:59:58.734085230Z I1204 11:59:58.734031 1 named_certificates.go:53] "Loaded SNI cert" index=0 certName="self-signed loopback" certDetail="\"apiserver-loopback-client@1764848711\" [serving] validServingFor=[apiserver-loopback-client] issuer=\"apiserver-loopback-client-ca@1764848711\" (2025-12-04 10:45:10 +0000 UTC to 2026-12-04 10:45:10 +0000 UTC (now=2025-12-04 11:59:58.734011907 +0000 UTC))" 2025-12-04T11:59:58.855092591Z I1204 11:59:58.854994 1 reflector.go:368] Caches populated for *v1.Secret from k8s.io/client-go/informers/factory.go:160 2025-12-04T11:59:58.858144008Z E1204 11:59:58.858054 1 base_controller.go:279] "Unhandled Error" err="OAuthServerRouteEndpointAccessibleController reconciliation failed: route \"openshift-authentication/oauth-openshift\": status does not have a valid host address" 2025-12-04T11:59:58.862205767Z E1204 11:59:58.862117 1 base_controller.go:279] "Unhandled Error" err="OAuthServerRouteEndpointAccessibleController reconciliation failed: route \"openshift-authentication/oauth-openshift\": status does not have a valid host address" 2025-12-04T11:59:59.042216438Z I1204 11:59:59.042001 1 request.go:700] Waited for 2.853751362s due to client-side throttling, not priority and fairness, request: GET:https://172.30.0.1:443/api/v1/namespaces/openshift-config-managed/configmaps?resourceVersion=11580 2025-12-04T11:59:59.094872246Z I1204 11:59:59.092692 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go/informers/factory.go:160 2025-12-04T11:59:59.103427067Z E1204 11:59:59.103035 1 base_controller.go:279] "Unhandled Error" err="OAuthServerRouteEndpointAccessibleController reconciliation failed: route \"openshift-authentication/oauth-openshift\": status does not have a valid host address" 2025-12-04T11:59:59.107544947Z E1204 11:59:59.107493 1 base_controller.go:279] "Unhandled Error" err="WellKnownReadyController reconciliation failed: failed to get oauth metadata from openshift-config-managed/oauth-openshift ConfigMap: configmap \"oauth-openshift\" not found (check authentication operator, it is supposed to create this)" 2025-12-04T11:59:59.108704414Z E1204 11:59:59.107830 1 base_controller.go:279] "Unhandled Error" err="OAuthServerRouteEndpointAccessibleController reconciliation failed: route \"openshift-authentication/oauth-openshift\": status does not have a valid host address" 2025-12-04T11:59:59.114691864Z E1204 11:59:59.114602 1 base_controller.go:279] "Unhandled Error" err="WellKnownReadyController reconciliation failed: failed to get oauth metadata from openshift-config-managed/oauth-openshift ConfigMap: configmap \"oauth-openshift\" not found (check authentication operator, it is supposed to create this)" 2025-12-04T11:59:59.158457870Z I1204 11:59:59.157968 1 reflector.go:368] Caches populated for *v1.Deployment from k8s.io/client-go/informers/factory.go:160 2025-12-04T11:59:59.174102165Z E1204 11:59:59.173984 1 base_controller.go:279] "Unhandled Error" err="OAuthServer-WorkloadWorkloadController reconciliation failed: waiting for the oauth-openshift route to contain an admitted ingress: no admitted ingress for route oauth-openshift in namespace openshift-authentication" 2025-12-04T11:59:59.443848798Z I1204 11:59:59.443680 1 reflector.go:368] Caches populated for *v1.Secret from k8s.io/client-go/informers/factory.go:160 2025-12-04T11:59:59.672740307Z I1204 11:59:59.672646 1 reflector.go:368] Caches populated for *v1.IngressController from github.com/openshift/client-go/operator/informers/externalversions/factory.go:125 2025-12-04T11:59:59.843667741Z I1204 11:59:59.843599 1 reflector.go:368] Caches populated for *v1.Secret from k8s.io/client-go/informers/factory.go:160 2025-12-04T12:00:00.241897103Z I1204 12:00:00.241798 1 request.go:700] Waited for 3.41635135s due to client-side throttling, not priority and fairness, request: GET:https://172.30.0.1:443/api/v1/namespaces/openshift-oauth-apiserver/services?resourceVersion=11582 2025-12-04T12:00:00.243782763Z I1204 12:00:00.243753 1 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 2025-12-04T12:00:00.444202061Z I1204 12:00:00.444133 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go/informers/factory.go:160 2025-12-04T12:00:00.644034209Z I1204 12:00:00.643938 1 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 2025-12-04T12:00:01.242452743Z I1204 12:00:01.242360 1 request.go:700] Waited for 3.357005836s due to client-side throttling, not priority and fairness, request: GET:https://172.30.0.1:443/api/v1/namespaces/openshift-oauth-apiserver/secrets/openshift-authenticator-certs 2025-12-04T12:00:02.151748821Z I1204 12:00:02.151620 1 reflector.go:368] Caches populated for *v1.Console from github.com/openshift/client-go/config/informers/externalversions/factory.go:125 2025-12-04T12:00:02.243702513Z I1204 12:00:02.243612 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go/informers/factory.go:160 2025-12-04T12:00:02.440967340Z I1204 12:00:02.440877 1 request.go:700] Waited for 4.218853616s due to client-side throttling, not priority and fairness, request: GET:https://172.30.0.1:443/api/v1/namespaces/openshift-etcd/endpoints?resourceVersion=11595 2025-12-04T12:00:02.443287734Z I1204 12:00:02.443245 1 reflector.go:368] Caches populated for *v1.Endpoints from k8s.io/client-go/informers/factory.go:160 2025-12-04T12:00:03.640955215Z I1204 12:00:03.640869 1 request.go:700] Waited for 2.594944715s due to client-side throttling, not priority and fairness, request: GET:https://172.30.0.1:443/api/v1/namespaces/openshift-config-managed/configmaps/oauth-serving-cert 2025-12-04T12:00:04.244738947Z I1204 12:00:04.243306 1 reflector.go:368] Caches populated for *v1.ConfigMap from k8s.io/client-go/informers/factory.go:160 2025-12-04T12:00:04.641299816Z I1204 12:00:04.641204 1 request.go:700] Waited for 2.396602653s due to client-side throttling, not priority and fairness, request: GET:https://172.30.0.1:443/api/v1/namespaces/openshift-oauth-apiserver/configmaps/audit 2025-12-04T12:00:05.641623097Z I1204 12:00:05.641554 1 request.go:700] Waited for 1.795592818s due to client-side throttling, not priority and fairness, request: GET:https://172.30.0.1:443/api/v1/namespaces/openshift-oauth-apiserver/secrets/openshift-authenticator-certs 2025-12-04T12:00:06.841706816Z I1204 12:00:06.841624 1 request.go:700] Waited for 1.401244929s due to client-side throttling, not priority and fairness, request: GET:https://172.30.0.1:443/api/v1/namespaces/openshift-authentication/secrets/v4-0-config-system-session 2025-12-04T12:00:07.582344173Z E1204 12:00:07.582263 1 base_controller.go:279] "Unhandled Error" err="OAuthServerRouteEndpointAccessibleController reconciliation failed: route \"openshift-authentication/oauth-openshift\": status does not have a valid host address" 2025-12-04T12:00:07.599765255Z E1204 12:00:07.599680 1 base_controller.go:279] "Unhandled Error" err="OAuthServerRouteEndpointAccessibleController reconciliation failed: route \"openshift-authentication/oauth-openshift\": status does not have a valid host address" 2025-12-04T12:00:08.642700575Z I1204 12:00:08.642562 1 request.go:700] Waited for 1.187822239s due to client-side throttling, not priority and fairness, request: GET:https://172.30.0.1:443/api/v1/namespaces/openshift-authentication/serviceaccounts/oauth-openshift 2025-12-04T12:00:09.152388098Z E1204 12:00:09.151416 1 base_controller.go:279] "Unhandled Error" err="OAuthServerServiceEndpointAccessibleController reconciliation failed: Get \"https://172.30.202.245:443/healthz\": dial tcp 172.30.202.245:443: connect: connection refused" 2025-12-04T12:00:09.841669438Z I1204 12:00:09.841549 1 request.go:700] Waited for 1.588152489s due to client-side throttling, not priority and fairness, request: GET:https://172.30.0.1:443/api/v1/namespaces/openshift-config/secrets/webhook-authentication-integrated-oauth 2025-12-04T12:00:11.041482047Z I1204 12:00:11.041384 1 request.go:700] Waited for 1.101510255s due to client-side throttling, not priority and fairness, request: GET:https://172.30.0.1:443/api/v1/namespaces/openshift-config/secrets/webhook-authentication-integrated-oauth 2025-12-04T12:00:11.168579853Z E1204 12:00:11.167900 1 base_controller.go:279] "Unhandled Error" err="OAuthServerServiceEndpointsEndpointAccessibleController reconciliation failed: oauth service endpoints are not ready" 2025-12-04T12:00:12.244123967Z I1204 12:00:12.243648 1 request.go:700] Waited for 1.000057953s due to client-side throttling, not priority and fairness, request: GET:https://172.30.0.1:443/api/v1/namespaces/openshift-oauth-apiserver/configmaps/trusted-ca-bundle 2025-12-04T12:00:16.112430340Z E1204 12:00:16.112316 1 base_controller.go:279] "Unhandled Error" err="OAuthServerRouteEndpointAccessibleController reconciliation failed: route \"openshift-authentication/oauth-openshift\": status does not have a valid host address" 2025-12-04T12:00:16.120911859Z E1204 12:00:16.120800 1 base_controller.go:279] "Unhandled Error" err="OAuthServerRouteEndpointAccessibleController reconciliation failed: route \"openshift-authentication/oauth-openshift\": status does not have a valid host address" 2025-12-04T12:00:16.732003473Z E1204 12:00:16.730872 1 base_controller.go:279] "Unhandled Error" err="OAuthServerRouteEndpointAccessibleController reconciliation failed: route \"openshift-authentication/oauth-openshift\": status does not have a valid host address" 2025-12-04T12:00:16.739940344Z E1204 12:00:16.739847 1 base_controller.go:279] "Unhandled Error" err="OAuthServerRouteEndpointAccessibleController reconciliation failed: route \"openshift-authentication/oauth-openshift\": status does not have a valid host address" 2025-12-04T12:00:16.816646063Z E1204 12:00:16.816485 1 base_controller.go:279] "Unhandled Error" err="OAuthServerRouteEndpointAccessibleController reconciliation failed: route \"openshift-authentication/oauth-openshift\": status does not have a valid host address" 2025-12-04T12:00:16.824325217Z E1204 12:00:16.824198 1 base_controller.go:279] "Unhandled Error" err="OAuthServerRouteEndpointAccessibleController reconciliation failed: route \"openshift-authentication/oauth-openshift\": status does not have a valid host address" 2025-12-04T12:00:17.892613891Z I1204 12:00:17.892520 1 reflector.go:368] Caches populated for *v1.OAuthClient from github.com/openshift/client-go/oauth/informers/externalversions/factory.go:125 2025-12-04T12:00:17.894608064Z I1204 12:00:17.894515 1 base_controller.go:82] Caches are synced for OAuthClientsController 2025-12-04T12:00:17.894608064Z I1204 12:00:17.894594 1 base_controller.go:119] Starting #1 worker of OAuthClientsController controller ... 2025-12-04T12:00:17.929897162Z E1204 12:00:17.929483 1 base_controller.go:279] "Unhandled Error" err="OAuthClientsController reconciliation failed: no ingress for host oauth-openshift.apps.sno.openstack.lab in route oauth-openshift in namespace openshift-authentication" 2025-12-04T12:00:17.935480599Z E1204 12:00:17.935310 1 base_controller.go:279] "Unhandled Error" err="OAuthServerServiceEndpointsEndpointAccessibleController reconciliation failed: oauth service endpoints are not ready" 2025-12-04T12:00:17.936144380Z I1204 12:00:17.936084 1 status_controller.go:225] clusteroperator/authentication diff {"status":{"conditions":[{"lastTransitionTime":"2025-12-04T11:56:30Z","message":"IngressStateEndpointsDegraded: No subsets found for the endpoints of oauth-server\nOAuthClientsControllerDegraded: no ingress for host oauth-openshift.apps.sno.openstack.lab in route oauth-openshift in namespace openshift-authentication\nOAuthServerDeploymentDegraded: waiting for the oauth-openshift route to contain an admitted ingress: no admitted ingress for route oauth-openshift in namespace openshift-authentication\nOAuthServerDeploymentDegraded: \nOAuthServerRouteEndpointAccessibleControllerDegraded: route \"openshift-authentication/oauth-openshift\": status does not have a valid host address\nOAuthServerServiceEndpointAccessibleControllerDegraded: Get \"https://172.30.202.245:443/healthz\": dial tcp 172.30.202.245:443: connect: connection refused\nOAuthServerServiceEndpointsEndpointAccessibleControllerDegraded: oauth service endpoints are not ready\nOAuthServerWorkloadDegraded: waiting for the oauth-openshift route to contain an admitted ingress: no admitted ingress for route oauth-openshift in namespace openshift-authentication\nOAuthServerWorkloadDegraded: \nWellKnownReadyControllerDegraded: failed to get oauth metadata from openshift-config-managed/oauth-openshift ConfigMap: configmap \"oauth-openshift\" not found (check authentication operator, it is supposed to create this)","reason":"IngressStateEndpoints_MissingSubsets::OAuthClientsController_SyncError::OAuthServerDeployment_PreconditionNotFulfilled::OAuthServerRouteEndpointAccessibleController_SyncError::OAuthServerServiceEndpointAccessibleController_SyncError::OAuthServerServiceEndpointsEndpointAccessibleController_SyncError::OAuthServerWorkload_PreconditionNotFulfilled::WellKnownReadyController_SyncError","status":"True","type":"Degraded"},{"lastTransitionTime":"2025-12-04T11:59:50Z","reason":"WellKnownReadyController","status":"True","type":"Progressing"},{"lastTransitionTime":"2025-12-04T11:37:45Z","message":"OAuthServerRouteEndpointAccessibleControllerAvailable: failed to retrieve route from cache: route.route.openshift.io \"oauth-openshift\" not found\nOAuthServerServiceEndpointAccessibleControllerAvailable: Get \"https://172.30.202.245:443/healthz\": dial tcp 172.30.202.245:443: connect: connection refused\nOAuthServerServiceEndpointsEndpointAccessibleControllerAvailable: endpoints \"oauth-openshift\" not found\nReadyIngressNodesAvailable: Authentication requires functional ingress which requires at least one schedulable and ready node. Got 0 worker nodes, 1 master nodes, 0 custom target nodes (none are schedulable or ready for ingress pods).\nWellKnownAvailable: The well-known endpoint is not yet available: failed to get oauth metadata from openshift-config-managed/oauth-openshift ConfigMap: configmap \"oauth-openshift\" not found (check authentication operator, it is supposed to create this)","reason":"OAuthServerDeployment_PreconditionNotFulfilled::OAuthServerRouteEndpointAccessibleController_ResourceNotFound::OAuthServerServiceEndpointAccessibleController_EndpointUnavailable::OAuthServerServiceEndpointsEndpointAccessibleController_ResourceNotFound::ReadyIngressNodes_NoReadyIngressNodes::WellKnown_NotReady","status":"False","type":"Available"},{"lastTransitionTime":"2025-12-04T11:37:45Z","message":"All is well","reason":"AsExpected","status":"True","type":"Upgradeable"},{"lastTransitionTime":"2025-12-04T11:37:45Z","reason":"NoData","status":"Unknown","type":"EvaluationConditionsDetected"}]}} 2025-12-04T12:00:17.936354497Z E1204 12:00:17.936315 1 base_controller.go:279] "Unhandled Error" err="OAuthServerRouteEndpointAccessibleController reconciliation failed: route \"openshift-authentication/oauth-openshift\": status does not have a valid host address" 2025-12-04T12:00:17.939179836Z E1204 12:00:17.939139 1 base_controller.go:279] "Unhandled Error" err="OAuthClientsController reconciliation failed: no ingress for host oauth-openshift.apps.sno.openstack.lab in route oauth-openshift in namespace openshift-authentication" 2025-12-04T12:00:17.943846084Z E1204 12:00:17.943760 1 base_controller.go:279] "Unhandled Error" err="OAuthServerServiceEndpointAccessibleController reconciliation failed: Get \"https://172.30.202.245:443/healthz\": dial tcp 172.30.202.245:443: connect: connection refused" 2025-12-04T12:00:17.954043096Z E1204 12:00:17.953949 1 base_controller.go:279] "Unhandled Error" err="OAuthClientsController reconciliation failed: no ingress for host oauth-openshift.apps.sno.openstack.lab in route oauth-openshift in namespace openshift-authentication" 2025-12-04T12:00:17.957418353Z I1204 12:00:17.957188 1 event.go:377] Event(v1.ObjectReference{Kind:"Deployment", Namespace:"openshift-authentication-operator", Name:"authentication-operator", UID:"", APIVersion:"", ResourceVersion:"", FieldPath:""}): type: 'Normal' reason: 'OperatorStatusChanged' Status for clusteroperator/authentication changed: Degraded message changed from "IngressStateEndpointsDegraded: No subsets found for the endpoints of oauth-server\nOAuthServerDeploymentDegraded: waiting for the oauth-openshift route to contain an admitted ingress: no admitted ingress for route oauth-openshift in namespace openshift-authentication\nOAuthServerDeploymentDegraded: \nOAuthServerRouteEndpointAccessibleControllerDegraded: route \"openshift-authentication/oauth-openshift\": status does not have a valid host address\nOAuthServerServiceEndpointAccessibleControllerDegraded: Get \"https://172.30.202.245:443/healthz\": dial tcp 172.30.202.245:443: connect: connection refused\nOAuthServerServiceEndpointsEndpointAccessibleControllerDegraded: oauth service endpoints are not ready\nOAuthServerWorkloadDegraded: waiting for the oauth-openshift route to contain an admitted ingress: no admitted ingress for route oauth-openshift in namespace openshift-authentication\nOAuthServerWorkloadDegraded: \nWellKnownReadyControllerDegraded: failed to get oauth metadata from openshift-config-managed/oauth-openshift ConfigMap: configmap \"oauth-openshift\" not found (check authentication operator, it is supposed to create this)" to "IngressStateEndpointsDegraded: No subsets found for the endpoints of oauth-server\nOAuthClientsControllerDegraded: no ingress for host oauth-openshift.apps.sno.openstack.lab in route oauth-openshift in namespace openshift-authentication\nOAuthServerDeploymentDegraded: waiting for the oauth-openshift route to contain an admitted ingress: no admitted ingress for route oauth-openshift in namespace openshift-authentication\nOAuthServerDeploymentDegraded: \nOAuthServerRouteEndpointAccessibleControllerDegraded: route \"openshift-authentication/oauth-openshift\": status does not have a valid host address\nOAuthServerServiceEndpointAccessibleControllerDegraded: Get \"https://172.30.202.245:443/healthz\": dial tcp 172.30.202.245:443: connect: connection refused\nOAuthServerServiceEndpointsEndpointAccessibleControllerDegraded: oauth service endpoints are not ready\nOAuthServerWorkloadDegraded: waiting for the oauth-openshift route to contain an admitted ingress: no admitted ingress for route oauth-openshift in namespace openshift-authentication\nOAuthServerWorkloadDegraded: \nWellKnownReadyControllerDegraded: failed to get oauth metadata from openshift-config-managed/oauth-openshift ConfigMap: configmap \"oauth-openshift\" not found (check authentication operator, it is supposed to create this)" 2025-12-04T12:00:17.980746182Z E1204 12:00:17.980658 1 base_controller.go:279] "Unhandled Error" err="OAuthClientsController reconciliation failed: no ingress for host oauth-openshift.apps.sno.openstack.lab in route oauth-openshift in namespace openshift-authentication" 2025-12-04T12:00:18.025743398Z E1204 12:00:18.025491 1 base_controller.go:279] "Unhandled Error" err="OAuthClientsController reconciliation failed: no ingress for host oauth-openshift.apps.sno.openstack.lab in route oauth-openshift in namespace openshift-authentication" 2025-12-04T12:00:18.110297875Z E1204 12:00:18.110176 1 base_controller.go:279] "Unhandled Error" err="OAuthClientsController reconciliation failed: no ingress for host oauth-openshift.apps.sno.openstack.lab in route oauth-openshift in namespace openshift-authentication" 2025-12-04T12:00:18.275733654Z E1204 12:00:18.275641 1 base_controller.go:279] "Unhandled Error" err="OAuthClientsController reconciliation failed: no ingress for host oauth-openshift.apps.sno.openstack.lab in route oauth-openshift in namespace openshift-authentication" 2025-12-04T12:00:18.509544680Z I1204 12:00:18.508996 1 request.go:700] Waited for 1.123354038s due to client-side throttling, not priority and fairness, request: GET:https://172.30.0.1:443/api/v1/namespaces/openshift-config/secrets/webhook-authentication-integrated-oauth 2025-12-04T12:00:18.601131940Z E1204 12:00:18.601021 1 base_controller.go:279] "Unhandled Error" err="OAuthClientsController reconciliation failed: no ingress for host oauth-openshift.apps.sno.openstack.lab in route oauth-openshift in namespace openshift-authentication" 2025-12-04T12:00:19.244877948Z E1204 12:00:19.244726 1 base_controller.go:279] "Unhandled Error" err="OAuthClientsController reconciliation failed: no ingress for host oauth-openshift.apps.sno.openstack.lab in route oauth-openshift in namespace openshift-authentication" 2025-12-04T12:00:19.509403057Z I1204 12:00:19.509301 1 request.go:700] Waited for 1.576368136s due to client-side throttling, not priority and fairness, request: GET:https://172.30.0.1:443/api/v1/namespaces/openshift-oauth-apiserver/configmaps/audit 2025-12-04T12:00:20.530468605Z E1204 12:00:20.530378 1 base_controller.go:279] "Unhandled Error" err="OAuthClientsController reconciliation failed: no ingress for host oauth-openshift.apps.sno.openstack.lab in route oauth-openshift in namespace openshift-authentication" 2025-12-04T12:00:20.709269017Z I1204 12:00:20.709188 1 request.go:700] Waited for 1.796146195s due to client-side throttling, not priority and fairness, request: GET:https://172.30.0.1:443/api/v1/namespaces/openshift-config-managed/configmaps/oauth-serving-cert 2025-12-04T12:00:23.094232560Z E1204 12:00:23.094133 1 base_controller.go:279] "Unhandled Error" err="OAuthClientsController reconciliation failed: no ingress for host oauth-openshift.apps.sno.openstack.lab in route oauth-openshift in namespace openshift-authentication" 2025-12-04T12:00:28.218747600Z E1204 12:00:28.218640 1 base_controller.go:279] "Unhandled Error" err="OAuthClientsController reconciliation failed: no ingress for host oauth-openshift.apps.sno.openstack.lab in route oauth-openshift in namespace openshift-authentication" 2025-12-04T12:00:38.462423956Z E1204 12:00:38.462313 1 base_controller.go:279] "Unhandled Error" err="OAuthClientsController reconciliation failed: no ingress for host oauth-openshift.apps.sno.openstack.lab in route oauth-openshift in namespace openshift-authentication" 2025-12-04T12:00:38.772020381Z E1204 12:00:38.771884 1 base_controller.go:279] "Unhandled Error" err="WellKnownReadyController reconciliation failed: failed to get oauth metadata from openshift-config-managed/oauth-openshift ConfigMap: configmap \"oauth-openshift\" not found (check authentication operator, it is supposed to create this)" 2025-12-04T12:00:38.774482209Z E1204 12:00:38.774404 1 base_controller.go:279] "Unhandled Error" err="OAuthServerRouteEndpointAccessibleController reconciliation failed: route \"openshift-authentication/oauth-openshift\": status does not have a valid host address" 2025-12-04T12:00:40.139028126Z E1204 12:00:40.138926 1 base_controller.go:279] "Unhandled Error" err="OAuthServer-WorkloadWorkloadController reconciliation failed: waiting for the oauth-openshift route to contain an admitted ingress: no admitted ingress for route oauth-openshift in namespace openshift-authentication" 2025-12-04T12:00:57.199035853Z E1204 12:00:57.198937 1 base_controller.go:279] "Unhandled Error" err="WellKnownReadyController reconciliation failed: failed to get oauth metadata from openshift-config-managed/oauth-openshift ConfigMap: configmap \"oauth-openshift\" not found (check authentication operator, it is supposed to create this)" 2025-12-04T12:00:58.947353215Z E1204 12:00:58.947256 1 base_controller.go:279] "Unhandled Error" err="OAuthClientsController reconciliation failed: no ingress for host oauth-openshift.apps.sno.openstack.lab in route oauth-openshift in namespace openshift-authentication" 2025-12-04T12:01:15.540587856Z E1204 12:01:15.540463 1 base_controller.go:279] "Unhandled Error" err="OAuthServerRouteEndpointAccessibleController reconciliation failed: route \"openshift-authentication/oauth-openshift\": status does not have a valid host address" 2025-12-04T12:01:20.935377694Z E1204 12:01:20.935255 1 base_controller.go:279] "Unhandled Error" err="OAuthServerServiceEndpointAccessibleController reconciliation failed: Get \"https://172.30.202.245:443/healthz\": dial tcp 172.30.202.245:443: connect: connection refused" 2025-12-04T12:01:32.810692870Z E1204 12:01:32.809934 1 base_controller.go:279] "Unhandled Error" err="OAuthServer-WorkloadWorkloadController reconciliation failed: waiting for the oauth-openshift route to contain an admitted ingress: no admitted ingress for route oauth-openshift in namespace openshift-authentication" 2025-12-04T12:01:39.756528411Z E1204 12:01:39.756440 1 base_controller.go:279] "Unhandled Error" err="OAuthClientsController reconciliation failed: no ingress for host oauth-openshift.apps.sno.openstack.lab in route oauth-openshift in namespace openshift-authentication" 2025-12-04T12:01:39.912876168Z E1204 12:01:39.911950 1 base_controller.go:279] "Unhandled Error" err="OAuthClientsController reconciliation failed: no ingress for host oauth-openshift.apps.sno.openstack.lab in route oauth-openshift in namespace openshift-authentication" 2025-12-04T12:01:43.122240808Z E1204 12:01:43.120665 1 base_controller.go:279] "Unhandled Error" err="OAuthServer-WorkloadWorkloadController reconciliation failed: waiting for the oauth-openshift route to contain an admitted ingress: no admitted ingress for route oauth-openshift in namespace openshift-authentication" 2025-12-04T12:01:59.850973726Z E1204 12:01:59.850907 1 base_controller.go:279] "Unhandled Error" err="OAuthServerServiceEndpointsEndpointAccessibleController reconciliation failed: oauth service endpoints are not ready" 2025-12-04T12:02:02.063351577Z E1204 12:02:02.063251 1 base_controller.go:279] "Unhandled Error" err="OAuthServer-WorkloadWorkloadController reconciliation failed: waiting for the oauth-openshift route to contain an admitted ingress: no admitted ingress for route oauth-openshift in namespace openshift-authentication" authentication-operator_previous.log0000640000000000000000000000336315114274105040110 0ustar0000000000000000config/pod/openshift-authentication-operator/logs/authentication-operator-6c968fdfdf-nrrfw2025-12-04T11:42:12.213149353Z Copying system trust bundle 2025-12-04T11:42:12.284328855Z W1204 11:42:12.284217 1 cmd.go:167] Unable to read initial content of "/tmp/terminate": open /tmp/terminate: no such file or directory 2025-12-04T11:42:12.284900975Z I1204 11:42:12.284847 1 cmd.go:253] Using service-serving-cert provided certificates 2025-12-04T11:42:12.284900975Z I1204 11:42:12.284876 1 leaderelection.go:121] The leader election gives 4 retries and allows for 30s of clock skew. The kube-apiserver downtime tolerance is 78s. Worst non-graceful lease acquisition is 2m43s. Worst graceful lease acquisition is {26s}. 2025-12-04T11:42:12.286265379Z I1204 11:42:12.286203 1 observer_polling.go:159] Starting file observer 2025-12-04T11:42:12.286283540Z I1204 11:42:12.286231 1 observer_polling.go:159] Starting file observer 2025-12-04T11:43:12.292393577Z W1204 11:43:12.292234 1 builder.go:272] unable to get owner reference (falling back to namespace): the server was unable to return a response in the time allotted, but may still be processing the request (get pods authentication-operator-6c968fdfdf-nrrfw) 2025-12-04T11:43:12.292515731Z I1204 11:43:12.292472 1 builder.go:304] cluster-authentication-operator version - 2025-12-04T11:43:12.293518227Z I1204 11:43:12.293459 1 dynamic_serving_content.go:116] "Loaded a new cert/key pair" name="serving-cert::/var/run/secrets/serving-cert/tls.crt::/var/run/secrets/serving-cert/tls.key" 2025-12-04T11:43:42.815579672Z F1204 11:43:42.815469 1 cmd.go:182] failed checking apiserver connectivity: Get "https://172.30.0.1:443/apis/coordination.k8s.io/v1/namespaces/openshift-authentication-operator/leases/cluster-authentication-operator-lock": context deadline exceeded config/pod/openshift-authentication-operator/authentication-operator-6c968fdfdf-nrrfw.json0000640000000000000000000002103615114274105030604 0ustar0000000000000000{"metadata":{"name":"authentication-operator-6c968fdfdf-nrrfw","generateName":"authentication-operator-6c968fdfdf-","namespace":"openshift-authentication-operator","uid":"42b3be0f-1d82-4a64-abb4-0118a6960efd","resourceVersion":"9874","creationTimestamp":"2025-12-04T11:35:49Z","labels":{"app":"authentication-operator","pod-template-hash":"6c968fdfdf"},"annotations":{"k8s.ovn.org/pod-networks":"{\"default\":{\"ip_addresses\":[\"10.128.0.11/23\"],\"mac_address\":\"0a:58:0a:80:00:0b\",\"gateway_ips\":[\"10.128.0.1\"],\"routes\":[{\"dest\":\"10.128.0.0/16\",\"nextHop\":\"10.128.0.1\"},{\"dest\":\"172.30.0.0/16\",\"nextHop\":\"10.128.0.1\"},{\"dest\":\"169.254.0.5/32\",\"nextHop\":\"10.128.0.1\"},{\"dest\":\"100.64.0.0/16\",\"nextHop\":\"10.128.0.1\"}],\"ip_address\":\"10.128.0.11/23\",\"gateway_ip\":\"10.128.0.1\",\"role\":\"primary\"}}","k8s.v1.cni.cncf.io/network-status":"[{\n \"name\": \"ovn-kubernetes\",\n \"interface\": \"eth0\",\n \"ips\": [\n \"10.128.0.11\"\n ],\n \"mac\": \"0a:58:0a:80:00:0b\",\n \"default\": true,\n \"dns\": {}\n}]","openshift.io/required-scc":"anyuid","openshift.io/scc":"anyuid"},"ownerReferences":[{"apiVersion":"apps/v1","kind":"ReplicaSet","name":"authentication-operator-6c968fdfdf","uid":"fae77f99-7c89-4ffd-aa39-d534d480ffce","controller":true,"blockOwnerDeletion":true}]},"spec":{"volumes":[{"name":"config","configMap":{"name":"authentication-operator-config","defaultMode":440}},{"name":"trusted-ca-bundle","configMap":{"name":"trusted-ca-bundle","defaultMode":420,"optional":true}},{"name":"serving-cert","secret":{"secretName":"serving-cert","defaultMode":420,"optional":true}},{"name":"service-ca-bundle","configMap":{"name":"service-ca-bundle","defaultMode":420,"optional":true}},{"name":"kube-api-access-c7xwd","projected":{"sources":[{"serviceAccountToken":{"expirationSeconds":3607,"path":"token"}},{"configMap":{"name":"kube-root-ca.crt","items":[{"key":"ca.crt","path":"ca.crt"}]}},{"downwardAPI":{"items":[{"path":"namespace","fieldRef":{"apiVersion":"v1","fieldPath":"metadata.namespace"}}]}},{"configMap":{"name":"openshift-service-ca.crt","items":[{"key":"service-ca.crt","path":"service-ca.crt"}]}}],"defaultMode":420}}],"containers":[{"name":"authentication-operator","image":"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e85850a4ae1a1e3ec2c590a4936d640882b6550124da22031c85b526afbf52df","command":["/bin/bash","-ec"],"args":["if [ -s /var/run/configmaps/trusted-ca-bundle/ca-bundle.crt ]; then\n echo \"Copying system trust bundle\"\n cp -f /var/run/configmaps/trusted-ca-bundle/ca-bundle.crt /etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem\nfi\nexec authentication-operator operator --config=/var/run/configmaps/config/operator-config.yaml --v=2 --terminate-on-files=/var/run/configmaps/trusted-ca-bundle/ca-bundle.crt --terminate-on-files=/tmp/terminate\n"],"env":[{"name":"IMAGE_OAUTH_SERVER","value":"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8860e00f858d1bca98344f21b5a5c4acc43c9c6eca8216582514021f0ab3cf7b"},{"name":"IMAGE_OAUTH_APISERVER","value":"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:91af633e585621630c40d14f188e37d36b44678d0a59e582d850bf8d593d3a0c"},{"name":"OPERATOR_IMAGE_VERSION","value":"4.18.29"},{"name":"OPERAND_OAUTH_SERVER_IMAGE_VERSION","value":"4.18.29_openshift"},{"name":"POD_NAME","valueFrom":{"fieldRef":{"apiVersion":"v1","fieldPath":"metadata.name"}}}],"resources":{"requests":{"cpu":"20m","memory":"200Mi"}},"volumeMounts":[{"name":"config","mountPath":"/var/run/configmaps/config"},{"name":"serving-cert","mountPath":"/var/run/secrets/serving-cert"},{"name":"trusted-ca-bundle","readOnly":true,"mountPath":"/var/run/configmaps/trusted-ca-bundle"},{"name":"service-ca-bundle","readOnly":true,"mountPath":"/var/run/configmaps/service-ca-bundle"},{"name":"kube-api-access-c7xwd","readOnly":true,"mountPath":"/var/run/secrets/kubernetes.io/serviceaccount"}],"livenessProbe":{"httpGet":{"path":"healthz","port":8443,"scheme":"HTTPS"},"initialDelaySeconds":30,"timeoutSeconds":1,"periodSeconds":10,"successThreshold":1,"failureThreshold":3},"terminationMessagePath":"/dev/termination-log","terminationMessagePolicy":"FallbackToLogsOnError","imagePullPolicy":"IfNotPresent","securityContext":{"capabilities":{"drop":["MKNOD"]},"runAsUser":0,"readOnlyRootFilesystem":false}}],"restartPolicy":"Always","terminationGracePeriodSeconds":30,"dnsPolicy":"ClusterFirst","nodeSelector":{"node-role.kubernetes.io/master":""},"serviceAccountName":"authentication-operator","serviceAccount":"authentication-operator","nodeName":"master-0","securityContext":{"seLinuxOptions":{"level":"s0:c16,c10"}},"schedulerName":"default-scheduler","tolerations":[{"key":"node-role.kubernetes.io/master","operator":"Exists","effect":"NoSchedule"},{"key":"node.kubernetes.io/unreachable","operator":"Exists","effect":"NoExecute","tolerationSeconds":120},{"key":"node.kubernetes.io/not-ready","operator":"Exists","effect":"NoExecute","tolerationSeconds":120},{"key":"node.kubernetes.io/memory-pressure","operator":"Exists","effect":"NoSchedule"}],"priorityClassName":"system-cluster-critical","priority":2000000000,"enableServiceLinks":true,"preemptionPolicy":"PreemptLowerPriority"},"status":{"phase":"Running","conditions":[{"type":"PodReadyToStartContainers","status":"True","lastProbeTime":null,"lastTransitionTime":"2025-12-04T11:37:43Z"},{"type":"Initialized","status":"True","lastProbeTime":null,"lastTransitionTime":"2025-12-04T11:37:30Z"},{"type":"Ready","status":"True","lastProbeTime":null,"lastTransitionTime":"2025-12-04T11:45:10Z"},{"type":"ContainersReady","status":"True","lastProbeTime":null,"lastTransitionTime":"2025-12-04T11:45:10Z"},{"type":"PodScheduled","status":"True","lastProbeTime":null,"lastTransitionTime":"2025-12-04T11:37:30Z"}],"hostIP":"192.168.32.10","hostIPs":[{"ip":"192.168.32.10"}],"podIP":"10.128.0.11","podIPs":[{"ip":"10.128.0.11"}],"startTime":"2025-12-04T11:37:30Z","containerStatuses":[{"name":"authentication-operator","state":{"running":{"startedAt":"2025-12-04T11:44:28Z"}},"lastState":{"terminated":{"exitCode":255,"reason":"Error","message":"Copying system trust bundle\nW1204 11:42:12.284217 1 cmd.go:167] Unable to read initial content of \"/tmp/terminate\": open /tmp/terminate: no such file or directory\nI1204 11:42:12.284847 1 cmd.go:253] Using service-serving-cert provided certificates\nI1204 11:42:12.284876 1 leaderelection.go:121] The leader election gives 4 retries and allows for 30s of clock skew. The kube-apiserver downtime tolerance is 78s. Worst non-graceful lease acquisition is 2m43s. Worst graceful lease acquisition is {26s}.\nI1204 11:42:12.286203 1 observer_polling.go:159] Starting file observer\nI1204 11:42:12.286231 1 observer_polling.go:159] Starting file observer\nW1204 11:43:12.292234 1 builder.go:272] unable to get owner reference (falling back to namespace): the server was unable to return a response in the time allotted, but may still be processing the request (get pods authentication-operator-6c968fdfdf-nrrfw)\nI1204 11:43:12.292472 1 builder.go:304] cluster-authentication-operator version -\nI1204 11:43:12.293459 1 dynamic_serving_content.go:116] \"Loaded a new cert/key pair\" name=\"serving-cert::/var/run/secrets/serving-cert/tls.crt::/var/run/secrets/serving-cert/tls.key\"\nF1204 11:43:42.815469 1 cmd.go:182] failed checking apiserver connectivity: Get \"https://172.30.0.1:443/apis/coordination.k8s.io/v1/namespaces/openshift-authentication-operator/leases/cluster-authentication-operator-lock\": context deadline exceeded\n","startedAt":"2025-12-04T11:42:12Z","finishedAt":"2025-12-04T11:43:42Z","containerID":"cri-o://a6d14a522051fb2206df06257821a6c013ab79cc060f6af18a92935eb6c12df9"}},"ready":true,"restartCount":3,"image":"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e85850a4ae1a1e3ec2c590a4936d640882b6550124da22031c85b526afbf52df","imageID":"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e85850a4ae1a1e3ec2c590a4936d640882b6550124da22031c85b526afbf52df","containerID":"cri-o://30a32ee2c91e00362356c25e062529920ef5ebdce171e46e07e4a515186d182b","started":true,"volumeMounts":[{"name":"config","mountPath":"/var/run/configmaps/config"},{"name":"serving-cert","mountPath":"/var/run/secrets/serving-cert"},{"name":"trusted-ca-bundle","mountPath":"/var/run/configmaps/trusted-ca-bundle","readOnly":true,"recursiveReadOnly":"Disabled"},{"name":"service-ca-bundle","mountPath":"/var/run/configmaps/service-ca-bundle","readOnly":true,"recursiveReadOnly":"Disabled"},{"name":"kube-api-access-c7xwd","mountPath":"/var/run/secrets/kubernetes.io/serviceaccount","readOnly":true,"recursiveReadOnly":"Disabled"}]}],"qosClass":"Burstable"}}events/openshift-operator-lifecycle-manager.json0000640000000000000000000006265015114274105020645 0ustar0000000000000000{"items":[{"namespace":"openshift-operator-lifecycle-manager","lastTimestamp":"2025-12-04T11:31:32Z","reason":"NoPods","message":"No matching pods found","type":"Normal"},{"namespace":"openshift-operator-lifecycle-manager","lastTimestamp":"2025-12-04T11:31:43Z","reason":"ScalingReplicaSet","message":"Scaled up replica set package-server-manager-67477646d4 to 1","type":"Normal"},{"namespace":"openshift-operator-lifecycle-manager","lastTimestamp":"2025-12-04T11:31:45Z","reason":"ScalingReplicaSet","message":"Scaled up replica set catalog-operator-fbc6455c4 to 1","type":"Normal"},{"namespace":"openshift-operator-lifecycle-manager","lastTimestamp":"2025-12-04T11:31:45Z","reason":"ScalingReplicaSet","message":"Scaled up replica set olm-operator-7cd7dbb44c to 1","type":"Normal"},{"namespace":"openshift-operator-lifecycle-manager","lastTimestamp":"2025-12-04T11:31:46Z","reason":"FailedCreate","message":"Error creating: pods \"package-server-manager-67477646d4-\" is forbidden: autoscaling.openshift.io/ManagementCPUsOverride the cluster does not have any nodes","type":"Warning"},{"namespace":"openshift-operator-lifecycle-manager","lastTimestamp":"2025-12-04T11:31:46Z","reason":"FailedCreate","message":"Error creating: pods \"olm-operator-7cd7dbb44c-\" is forbidden: autoscaling.openshift.io/ManagementCPUsOverride the cluster does not have any nodes","type":"Warning"},{"namespace":"openshift-operator-lifecycle-manager","lastTimestamp":"2025-12-04T11:31:46Z","reason":"FailedCreate","message":"Error creating: pods \"catalog-operator-fbc6455c4-\" is forbidden: autoscaling.openshift.io/ManagementCPUsOverride the cluster does not have any nodes","type":"Warning"},{"namespace":"openshift-operator-lifecycle-manager","lastTimestamp":"2025-12-04T11:35:47Z","reason":"NoPods","message":"No matching pods found","type":"Normal"},{"namespace":"openshift-operator-lifecycle-manager","lastTimestamp":"2025-12-04T11:35:47Z","reason":"FailedCreate","message":"Error creating: pods \"catalog-operator-fbc6455c4-\" is forbidden: autoscaling.openshift.io/ManagementCPUsOverride the cluster does not have any nodes","type":"Warning"},{"namespace":"openshift-operator-lifecycle-manager","lastTimestamp":"2025-12-04T11:35:48Z","reason":"FailedCreate","message":"Error creating: pods \"package-server-manager-67477646d4-\" is forbidden: autoscaling.openshift.io/ManagementCPUsOverride the cluster does not have any nodes","type":"Warning"},{"namespace":"openshift-operator-lifecycle-manager","lastTimestamp":"2025-12-04T11:35:48Z","reason":"FailedCreate","message":"Error creating: pods \"olm-operator-7cd7dbb44c-\" is forbidden: autoscaling.openshift.io/ManagementCPUsOverride the cluster does not have any nodes","type":"Warning"},{"namespace":"openshift-operator-lifecycle-manager","lastTimestamp":"2025-12-04T11:35:50Z","reason":"FailedScheduling","message":"0/1 nodes are available: 1 node(s) had untolerated taint {node.kubernetes.io/not-ready: }. preemption: 0/1 nodes are available: 1 Preemption is not helpful for scheduling.","type":"Warning"},{"namespace":"openshift-operator-lifecycle-manager","lastTimestamp":"2025-12-04T11:35:50Z","reason":"SuccessfulCreate","message":"Created pod: olm-operator-7cd7dbb44c-vzj4q","type":"Normal"},{"namespace":"openshift-operator-lifecycle-manager","lastTimestamp":"2025-12-04T11:35:50Z","reason":"FailedScheduling","message":"0/1 nodes are available: 1 node(s) had untolerated taint {node.kubernetes.io/not-ready: }. preemption: 0/1 nodes are available: 1 Preemption is not helpful for scheduling.","type":"Warning"},{"namespace":"openshift-operator-lifecycle-manager","lastTimestamp":"2025-12-04T11:35:50Z","reason":"FailedScheduling","message":"0/1 nodes are available: 1 node(s) had untolerated taint {node.kubernetes.io/not-ready: }. preemption: 0/1 nodes are available: 1 Preemption is not helpful for scheduling.","type":"Warning"},{"namespace":"openshift-operator-lifecycle-manager","lastTimestamp":"2025-12-04T11:35:50Z","reason":"SuccessfulCreate","message":"Created pod: catalog-operator-fbc6455c4-5m8ll","type":"Normal"},{"namespace":"openshift-operator-lifecycle-manager","lastTimestamp":"2025-12-04T11:35:50Z","reason":"SuccessfulCreate","message":"Created pod: package-server-manager-67477646d4-7hndf","type":"Normal"},{"namespace":"openshift-operator-lifecycle-manager","lastTimestamp":"2025-12-04T11:37:30Z","reason":"Scheduled","message":"Successfully assigned openshift-operator-lifecycle-manager/catalog-operator-fbc6455c4-5m8ll to master-0","type":"Normal"},{"namespace":"openshift-operator-lifecycle-manager","lastTimestamp":"2025-12-04T11:37:30Z","reason":"Scheduled","message":"Successfully assigned openshift-operator-lifecycle-manager/package-server-manager-67477646d4-7hndf to master-0","type":"Normal"},{"namespace":"openshift-operator-lifecycle-manager","lastTimestamp":"2025-12-04T11:37:30Z","reason":"Scheduled","message":"Successfully assigned openshift-operator-lifecycle-manager/olm-operator-7cd7dbb44c-vzj4q to master-0","type":"Normal"},{"namespace":"openshift-operator-lifecycle-manager","lastTimestamp":"2025-12-04T11:37:38Z","reason":"FailedMount","message":"MountVolume.SetUp failed for volume \"srv-cert\" : secret \"catalog-operator-serving-cert\" not found","type":"Warning"},{"namespace":"openshift-operator-lifecycle-manager","lastTimestamp":"2025-12-04T11:37:38Z","reason":"FailedMount","message":"MountVolume.SetUp failed for volume \"srv-cert\" : secret \"olm-operator-serving-cert\" not found","type":"Warning"},{"namespace":"openshift-operator-lifecycle-manager","lastTimestamp":"2025-12-04T11:37:38Z","reason":"FailedMount","message":"MountVolume.SetUp failed for volume \"package-server-manager-serving-cert\" : secret \"package-server-manager-serving-cert\" not found","type":"Warning"},{"namespace":"openshift-operator-lifecycle-manager","lastTimestamp":"2025-12-04T11:38:02Z","reason":"FailedMount","message":"MountVolume.SetUp failed for volume \"srv-cert\" : secret \"olm-operator-serving-cert\" not found","type":"Warning"},{"namespace":"openshift-operator-lifecycle-manager","lastTimestamp":"2025-12-04T11:38:02Z","reason":"FailedMount","message":"MountVolume.SetUp failed for volume \"srv-cert\" : secret \"catalog-operator-serving-cert\" not found","type":"Warning"},{"namespace":"openshift-operator-lifecycle-manager","lastTimestamp":"2025-12-04T11:38:02Z","reason":"FailedMount","message":"MountVolume.SetUp failed for volume \"package-server-manager-serving-cert\" : secret \"package-server-manager-serving-cert\" not found","type":"Warning"},{"namespace":"openshift-operator-lifecycle-manager","lastTimestamp":"2025-12-04T11:38:19Z","reason":"AddedInterface","message":"Add eth0 [10.128.0.22/23] from ovn-kubernetes","type":"Normal"},{"namespace":"openshift-operator-lifecycle-manager","lastTimestamp":"2025-12-04T11:38:19Z","reason":"AddedInterface","message":"Add eth0 [10.128.0.13/23] from ovn-kubernetes","type":"Normal"},{"namespace":"openshift-operator-lifecycle-manager","lastTimestamp":"2025-12-04T11:38:19Z","reason":"AddedInterface","message":"Add eth0 [10.128.0.15/23] from ovn-kubernetes","type":"Normal"},{"namespace":"openshift-operator-lifecycle-manager","lastTimestamp":"2025-12-04T11:38:19Z","reason":"Pulling","message":"Pulling image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f1ca78c423f43f89a0411e40393642f64e4f8df9e5f61c25e31047c4cce170f9\"","type":"Normal"},{"namespace":"openshift-operator-lifecycle-manager","lastTimestamp":"2025-12-04T11:38:20Z","reason":"Started","message":"Started container kube-rbac-proxy","type":"Normal"},{"namespace":"openshift-operator-lifecycle-manager","lastTimestamp":"2025-12-04T11:38:20Z","reason":"Pulled","message":"Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69ffd8f8dcceedc2d6eb306cea33f8beabc1be1308cd5f4ee8b9a8e3eab9843\" already present on machine","type":"Normal"},{"namespace":"openshift-operator-lifecycle-manager","lastTimestamp":"2025-12-04T11:38:20Z","reason":"Created","message":"Created container: kube-rbac-proxy","type":"Normal"},{"namespace":"openshift-operator-lifecycle-manager","lastTimestamp":"2025-12-04T11:38:20Z","reason":"Pulling","message":"Pulling image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f1ca78c423f43f89a0411e40393642f64e4f8df9e5f61c25e31047c4cce170f9\"","type":"Normal"},{"namespace":"openshift-operator-lifecycle-manager","lastTimestamp":"2025-12-04T11:38:20Z","reason":"Pulling","message":"Pulling image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f1ca78c423f43f89a0411e40393642f64e4f8df9e5f61c25e31047c4cce170f9\"","type":"Normal"},{"namespace":"openshift-operator-lifecycle-manager","lastTimestamp":"2025-12-04T11:38:32Z","reason":"Pulled","message":"Successfully pulled image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f1ca78c423f43f89a0411e40393642f64e4f8df9e5f61c25e31047c4cce170f9\" in 12.198s (12.198s including waiting). Image size: 857069957 bytes.","type":"Normal"},{"namespace":"openshift-operator-lifecycle-manager","lastTimestamp":"2025-12-04T11:38:32Z","reason":"Pulled","message":"Successfully pulled image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f1ca78c423f43f89a0411e40393642f64e4f8df9e5f61c25e31047c4cce170f9\" in 12.362s (12.362s including waiting). Image size: 857069957 bytes.","type":"Normal"},{"namespace":"openshift-operator-lifecycle-manager","lastTimestamp":"2025-12-04T11:38:32Z","reason":"Created","message":"Created container: catalog-operator","type":"Normal"},{"namespace":"openshift-operator-lifecycle-manager","lastTimestamp":"2025-12-04T11:38:32Z","reason":"Created","message":"Created container: olm-operator","type":"Normal"},{"namespace":"openshift-operator-lifecycle-manager","lastTimestamp":"2025-12-04T11:38:32Z","reason":"Pulled","message":"Successfully pulled image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f1ca78c423f43f89a0411e40393642f64e4f8df9e5f61c25e31047c4cce170f9\" in 12.694s (12.694s including waiting). Image size: 857069957 bytes.","type":"Normal"},{"namespace":"openshift-operator-lifecycle-manager","lastTimestamp":"2025-12-04T11:38:33Z","reason":"Started","message":"Started container catalog-operator","type":"Normal"},{"namespace":"openshift-operator-lifecycle-manager","lastTimestamp":"2025-12-04T11:38:33Z","reason":"LeaderElection","message":"package-server-manager-67477646d4-7hndf_82aab290-170a-44ef-abf0-39b8795ad511 became leader","type":"Normal"},{"namespace":"openshift-operator-lifecycle-manager","lastTimestamp":"2025-12-04T11:38:33Z","reason":"Started","message":"Started container olm-operator","type":"Normal"},{"namespace":"openshift-operator-lifecycle-manager","lastTimestamp":"2025-12-04T11:38:35Z","reason":"NoOperatorGroup","message":"csv in namespace with no operatorgroups","type":"Warning"},{"namespace":"openshift-operator-lifecycle-manager","lastTimestamp":"2025-12-04T11:38:44Z","reason":"Scheduled","message":"Successfully assigned openshift-operator-lifecycle-manager/packageserver-675f5c767c-mtdrq to master-0","type":"Normal"},{"namespace":"openshift-operator-lifecycle-manager","lastTimestamp":"2025-12-04T11:38:44Z","reason":"RequirementsUnknown","message":"InstallModes now support target namespaces","type":"Normal"},{"namespace":"openshift-operator-lifecycle-manager","lastTimestamp":"2025-12-04T11:38:44Z","reason":"SuccessfulCreate","message":"Created pod: packageserver-675f5c767c-mtdrq","type":"Normal"},{"namespace":"openshift-operator-lifecycle-manager","lastTimestamp":"2025-12-04T11:38:44Z","reason":"ScalingReplicaSet","message":"Scaled up replica set packageserver-675f5c767c to 1","type":"Normal"},{"namespace":"openshift-operator-lifecycle-manager","lastTimestamp":"2025-12-04T11:40:22Z","reason":"FailedCreatePodSandBox","message":"Failed to create pod sandbox: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_packageserver-675f5c767c-mtdrq_openshift-operator-lifecycle-manager_855d7874-16b1-47d0-82f6-d2b0c89b9a84_0(896aef82e1d6cb28b4d2159f1388c189773aa632f6eace1e5d447c4fed3a97a7): error adding pod openshift-operator-lifecycle-manager_packageserver-675f5c767c-mtdrq to CNI network \"multus-cni-network\": plugin type=\"multus-shim\" name=\"multus-cni-network\" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:\"896aef82e1d6cb28b4d2159f1388c189773aa632f6eace1e5d447c4fed3a97a7\" Netns:\"/var/run/netns/8aec4def-e56e-4b8f-ae8d-b88220ffeb08\" IfName:\"eth0\" Args:\"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-operator-lifecycle-manager;K8S_POD_NAME=packageserver-675f5c767c-mtdrq;K8S_POD_INFRA_CONTAINER_ID=896aef82e1d6cb28b4d2159f1388c189773aa632f6eace1e5d447c4fed3a97a7;K8S_POD_UID=855d7874-16b1-47d0-82f6-d2b0c89b9a84\" Path:\"\" ERRORED: error configuring pod [openshift-operator-lifecycle-manager/packageserver-675f5c767c-mtdrq] networking: Multus: [openshift-operator-lifecycle-manager/packageserver-675f5c767c-mtdrq/855d7874-16b1-47d0-82f6-d2b0c89b9a84]: error setting the networks status: SetPodNetworkStatusAnnotation: failed to update the pod packageserver-675f5c767c-mtdrq in out of cluster comm: SetNetworkStatus: failed to update the pod packageserver-675f5c767c-mtdrq in out of cluster comm: status update failed for pod /: Get \"https://api-int.sno.openstack.lab:6443/api/v1/namespaces/openshift-operator-lifecycle-manager/pods/packageserver-675f5c767c-mtdrq?timeout=1m0s\": context deadline exceeded\n': StdinData: {\"binDir\":\"/var/lib/cni/bin\",\"clusterNetwork\":\"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf\",\"cniVersion\":\"0.3.1\",\"daemonSocketDir\":\"/run/multus/socket\",\"globalNamespaces\":\"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv\",\"logLevel\":\"verbose\",\"logToStderr\":true,\"name\":\"multus-cni-network\",\"namespaceIsolation\":true,\"type\":\"multus-shim\"}","type":"Warning"},{"namespace":"openshift-operator-lifecycle-manager","lastTimestamp":"2025-12-04T11:41:23Z","reason":"FailedCreatePodSandBox","message":"Failed to create pod sandbox: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_packageserver-675f5c767c-mtdrq_openshift-operator-lifecycle-manager_855d7874-16b1-47d0-82f6-d2b0c89b9a84_0(b5269202f491395be1819acda94362a4aa1e2a9aed4f6562916b06241919d0f4): error adding pod openshift-operator-lifecycle-manager_packageserver-675f5c767c-mtdrq to CNI network \"multus-cni-network\": plugin type=\"multus-shim\" name=\"multus-cni-network\" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:\"b5269202f491395be1819acda94362a4aa1e2a9aed4f6562916b06241919d0f4\" Netns:\"/var/run/netns/59d1358a-5227-419d-89e4-6be9839ac080\" IfName:\"eth0\" Args:\"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-operator-lifecycle-manager;K8S_POD_NAME=packageserver-675f5c767c-mtdrq;K8S_POD_INFRA_CONTAINER_ID=b5269202f491395be1819acda94362a4aa1e2a9aed4f6562916b06241919d0f4;K8S_POD_UID=855d7874-16b1-47d0-82f6-d2b0c89b9a84\" Path:\"\" ERRORED: error configuring pod [openshift-operator-lifecycle-manager/packageserver-675f5c767c-mtdrq] networking: Multus: [openshift-operator-lifecycle-manager/packageserver-675f5c767c-mtdrq/855d7874-16b1-47d0-82f6-d2b0c89b9a84]: error setting the networks status: SetPodNetworkStatusAnnotation: failed to update the pod packageserver-675f5c767c-mtdrq in out of cluster comm: SetNetworkStatus: failed to update the pod packageserver-675f5c767c-mtdrq in out of cluster comm: status update failed for pod /: Get \"https://api-int.sno.openstack.lab:6443/api/v1/namespaces/openshift-operator-lifecycle-manager/pods/packageserver-675f5c767c-mtdrq?timeout=1m0s\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)\n': StdinData: {\"binDir\":\"/var/lib/cni/bin\",\"clusterNetwork\":\"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf\",\"cniVersion\":\"0.3.1\",\"daemonSocketDir\":\"/run/multus/socket\",\"globalNamespaces\":\"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv\",\"logLevel\":\"verbose\",\"logToStderr\":true,\"name\":\"multus-cni-network\",\"namespaceIsolation\":true,\"type\":\"multus-shim\"}","type":"Warning"},{"namespace":"openshift-operator-lifecycle-manager","lastTimestamp":"2025-12-04T11:43:12Z","reason":"FailedCreatePodSandBox","message":"Failed to create pod sandbox: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_packageserver-675f5c767c-mtdrq_openshift-operator-lifecycle-manager_855d7874-16b1-47d0-82f6-d2b0c89b9a84_0(50d5ddf79bd5de442fb43ec994762ef41de345f87e9c87e14eaf6c0f2b22e980): error adding pod openshift-operator-lifecycle-manager_packageserver-675f5c767c-mtdrq to CNI network \"multus-cni-network\": plugin type=\"multus-shim\" name=\"multus-cni-network\" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:\"50d5ddf79bd5de442fb43ec994762ef41de345f87e9c87e14eaf6c0f2b22e980\" Netns:\"/var/run/netns/9708ead0-bfd0-4a87-9250-4e5efff45dab\" IfName:\"eth0\" Args:\"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-operator-lifecycle-manager;K8S_POD_NAME=packageserver-675f5c767c-mtdrq;K8S_POD_INFRA_CONTAINER_ID=50d5ddf79bd5de442fb43ec994762ef41de345f87e9c87e14eaf6c0f2b22e980;K8S_POD_UID=855d7874-16b1-47d0-82f6-d2b0c89b9a84\" Path:\"\" ERRORED: error configuring pod [openshift-operator-lifecycle-manager/packageserver-675f5c767c-mtdrq] networking: Multus: [openshift-operator-lifecycle-manager/packageserver-675f5c767c-mtdrq/855d7874-16b1-47d0-82f6-d2b0c89b9a84]: error setting the networks status: SetPodNetworkStatusAnnotation: failed to update the pod packageserver-675f5c767c-mtdrq in out of cluster comm: SetNetworkStatus: failed to update the pod packageserver-675f5c767c-mtdrq in out of cluster comm: status update failed for pod /: Get \"https://api-int.sno.openstack.lab:6443/api/v1/namespaces/openshift-operator-lifecycle-manager/pods/packageserver-675f5c767c-mtdrq?timeout=1m0s\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)\n': StdinData: {\"binDir\":\"/var/lib/cni/bin\",\"clusterNetwork\":\"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf\",\"cniVersion\":\"0.3.1\",\"daemonSocketDir\":\"/run/multus/socket\",\"globalNamespaces\":\"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv\",\"logLevel\":\"verbose\",\"logToStderr\":true,\"name\":\"multus-cni-network\",\"namespaceIsolation\":true,\"type\":\"multus-shim\"}","type":"Warning"},{"namespace":"openshift-operator-lifecycle-manager","lastTimestamp":"2025-12-04T11:44:18Z","reason":"ProbeError","message":"Liveness probe error: Get \"http://10.128.0.13:8080/healthz\": dial tcp 10.128.0.13:8080: connect: connection refused\nbody: \n","type":"Warning"},{"namespace":"openshift-operator-lifecycle-manager","lastTimestamp":"2025-12-04T11:44:18Z","reason":"Unhealthy","message":"Readiness probe failed: Get \"http://10.128.0.13:8080/healthz\": dial tcp 10.128.0.13:8080: connect: connection refused","type":"Warning"},{"namespace":"openshift-operator-lifecycle-manager","lastTimestamp":"2025-12-04T11:44:18Z","reason":"Unhealthy","message":"Liveness probe failed: Get \"http://10.128.0.13:8080/healthz\": dial tcp 10.128.0.13:8080: connect: connection refused","type":"Warning"},{"namespace":"openshift-operator-lifecycle-manager","lastTimestamp":"2025-12-04T11:44:18Z","reason":"ProbeError","message":"Readiness probe error: Get \"http://10.128.0.13:8080/healthz\": dial tcp 10.128.0.13:8080: connect: connection refused\nbody: \n","type":"Warning"},{"namespace":"openshift-operator-lifecycle-manager","lastTimestamp":"2025-12-04T11:44:28Z","reason":"Started","message":"Started container package-server-manager","type":"Normal"},{"namespace":"openshift-operator-lifecycle-manager","lastTimestamp":"2025-12-04T11:44:54Z","reason":"FailedCreatePodSandBox","message":"Failed to create pod sandbox: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_packageserver-675f5c767c-mtdrq_openshift-operator-lifecycle-manager_855d7874-16b1-47d0-82f6-d2b0c89b9a84_0(08bff72cbcb288755b606d271a90590d5de4969e52f24b14dfba197207176aab): error adding pod openshift-operator-lifecycle-manager_packageserver-675f5c767c-mtdrq to CNI network \"multus-cni-network\": plugin type=\"multus-shim\" name=\"multus-cni-network\" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:\"08bff72cbcb288755b606d271a90590d5de4969e52f24b14dfba197207176aab\" Netns:\"/var/run/netns/a579b1af-3832-4c4e-8db5-2e91b975a454\" IfName:\"eth0\" Args:\"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-operator-lifecycle-manager;K8S_POD_NAME=packageserver-675f5c767c-mtdrq;K8S_POD_INFRA_CONTAINER_ID=08bff72cbcb288755b606d271a90590d5de4969e52f24b14dfba197207176aab;K8S_POD_UID=855d7874-16b1-47d0-82f6-d2b0c89b9a84\" Path:\"\" ERRORED: error configuring pod [openshift-operator-lifecycle-manager/packageserver-675f5c767c-mtdrq] networking: Multus: [openshift-operator-lifecycle-manager/packageserver-675f5c767c-mtdrq/855d7874-16b1-47d0-82f6-d2b0c89b9a84]: error setting the networks status: SetPodNetworkStatusAnnotation: failed to update the pod packageserver-675f5c767c-mtdrq in out of cluster comm: SetNetworkStatus: failed to update the pod packageserver-675f5c767c-mtdrq in out of cluster comm: status update failed for pod /: the server was unable to return a response in the time allotted, but may still be processing the request (get pods packageserver-675f5c767c-mtdrq)\n': StdinData: {\"binDir\":\"/var/lib/cni/bin\",\"clusterNetwork\":\"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf\",\"cniVersion\":\"0.3.1\",\"daemonSocketDir\":\"/run/multus/socket\",\"globalNamespaces\":\"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv\",\"logLevel\":\"verbose\",\"logToStderr\":true,\"name\":\"multus-cni-network\",\"namespaceIsolation\":true,\"type\":\"multus-shim\"}","type":"Warning"},{"namespace":"openshift-operator-lifecycle-manager","lastTimestamp":"2025-12-04T11:45:11Z","reason":"Created","message":"Created container: packageserver","type":"Normal"},{"namespace":"openshift-operator-lifecycle-manager","lastTimestamp":"2025-12-04T11:45:11Z","reason":"Pulled","message":"Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f1ca78c423f43f89a0411e40393642f64e4f8df9e5f61c25e31047c4cce170f9\" already present on machine","type":"Normal"},{"namespace":"openshift-operator-lifecycle-manager","lastTimestamp":"2025-12-04T11:45:11Z","reason":"Started","message":"Started container packageserver","type":"Normal"},{"namespace":"openshift-operator-lifecycle-manager","lastTimestamp":"2025-12-04T11:45:11Z","reason":"AddedInterface","message":"Add eth0 [10.128.0.64/23] from ovn-kubernetes","type":"Normal"},{"namespace":"openshift-operator-lifecycle-manager","lastTimestamp":"2025-12-04T11:45:13Z","reason":"ProbeError","message":"Readiness probe error: Get \"https://10.128.0.64:5443/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)\nbody: \n","type":"Warning"},{"namespace":"openshift-operator-lifecycle-manager","lastTimestamp":"2025-12-04T11:45:13Z","reason":"Unhealthy","message":"Readiness probe failed: Get \"https://10.128.0.64:5443/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)","type":"Warning"},{"namespace":"openshift-operator-lifecycle-manager","lastTimestamp":"2025-12-04T11:45:15Z","reason":"ProbeError","message":"Readiness probe error: Get \"https://10.128.0.64:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)\nbody: \n","type":"Warning"},{"namespace":"openshift-operator-lifecycle-manager","lastTimestamp":"2025-12-04T11:45:15Z","reason":"Unhealthy","message":"Readiness probe failed: Get \"https://10.128.0.64:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)","type":"Warning"},{"namespace":"openshift-operator-lifecycle-manager","lastTimestamp":"2025-12-04T11:49:29Z","reason":"LeaderElection","message":"package-server-manager-67477646d4-7hndf_5a767bc3-9a75-405d-8427-31679e2dfc65 became leader","type":"Normal"},{"namespace":"openshift-operator-lifecycle-manager","lastTimestamp":"2025-12-04T11:53:38Z","reason":"BackOff","message":"Back-off restarting failed container package-server-manager in pod package-server-manager-67477646d4-7hndf_openshift-operator-lifecycle-manager(72faf6d6-e8ca-43d1-b93e-67c11f8d3b46)","type":"Warning"},{"namespace":"openshift-operator-lifecycle-manager","lastTimestamp":"2025-12-04T11:53:54Z","reason":"Pulled","message":"Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f1ca78c423f43f89a0411e40393642f64e4f8df9e5f61c25e31047c4cce170f9\" already present on machine","type":"Normal"},{"namespace":"openshift-operator-lifecycle-manager","lastTimestamp":"2025-12-04T11:53:54Z","reason":"Created","message":"Created container: package-server-manager","type":"Normal"},{"namespace":"openshift-operator-lifecycle-manager","lastTimestamp":"2025-12-04T11:58:24Z","reason":"LeaderElection","message":"package-server-manager-67477646d4-7hndf_efa69d16-9250-446b-9dbd-9d804226ad9f became leader","type":"Normal"},{"namespace":"openshift-operator-lifecycle-manager","lastTimestamp":"2025-12-04T12:00:59Z","reason":"AllRequirementsMet","message":"all requirements found, attempting install","type":"Normal"},{"namespace":"openshift-operator-lifecycle-manager","lastTimestamp":"2025-12-04T12:00:59Z","reason":"InstallSucceeded","message":"waiting for install components to report healthy","type":"Normal"},{"namespace":"openshift-operator-lifecycle-manager","lastTimestamp":"2025-12-04T12:00:59Z","reason":"InstallWaiting","message":"apiServices not installed","type":"Normal"},{"namespace":"openshift-operator-lifecycle-manager","lastTimestamp":"2025-12-04T12:00:59Z","reason":"InstallCheckFailed","message":"install timeout","type":"Warning"},{"namespace":"openshift-operator-lifecycle-manager","lastTimestamp":"2025-12-04T12:00:59Z","reason":"NeedsReinstall","message":"apiServices not installed","type":"Normal"}]}events/openshift-route-controller-manager.json0000640000000000000000000006244715114274105020400 0ustar0000000000000000{"items":[{"namespace":"openshift-route-controller-manager","lastTimestamp":"2025-12-04T11:37:46Z","reason":"Scheduled","message":"Successfully assigned openshift-route-controller-manager/route-controller-manager-864c894b4d-xbpkz to master-0","type":"Normal"},{"namespace":"openshift-route-controller-manager","lastTimestamp":"2025-12-04T11:37:46Z","reason":"SuccessfulCreate","message":"Created pod: route-controller-manager-864c894b4d-xbpkz","type":"Normal"},{"namespace":"openshift-route-controller-manager","lastTimestamp":"2025-12-04T11:37:46Z","reason":"ScalingReplicaSet","message":"Scaled up replica set route-controller-manager-864c894b4d to 1","type":"Normal"},{"namespace":"openshift-route-controller-manager","lastTimestamp":"2025-12-04T11:37:47Z","reason":"FailedMount","message":"MountVolume.SetUp failed for volume \"config\" : configmap \"config\" not found","type":"Warning"},{"namespace":"openshift-route-controller-manager","lastTimestamp":"2025-12-04T11:37:48Z","reason":"ScalingReplicaSet","message":"Scaled up replica set route-controller-manager-5f845c897b to 1 from 0","type":"Normal"},{"namespace":"openshift-route-controller-manager","lastTimestamp":"2025-12-04T11:37:48Z","reason":"FailedMount","message":"MountVolume.SetUp failed for volume \"serving-cert\" : secret \"serving-cert\" not found","type":"Warning"},{"namespace":"openshift-route-controller-manager","lastTimestamp":"2025-12-04T11:37:48Z","reason":"FailedMount","message":"MountVolume.SetUp failed for volume \"client-ca\" : configmap \"client-ca\" not found","type":"Warning"},{"namespace":"openshift-route-controller-manager","lastTimestamp":"2025-12-04T11:37:48Z","reason":"SuccessfulCreate","message":"Created pod: route-controller-manager-5f845c897b-hhhv6","type":"Normal"},{"namespace":"openshift-route-controller-manager","lastTimestamp":"2025-12-04T11:37:48Z","reason":"ScalingReplicaSet","message":"Scaled down replica set route-controller-manager-864c894b4d to 0 from 1","type":"Normal"},{"namespace":"openshift-route-controller-manager","lastTimestamp":"2025-12-04T11:37:48Z","reason":"SuccessfulDelete","message":"Deleted pod: route-controller-manager-864c894b4d-xbpkz","type":"Normal"},{"namespace":"openshift-route-controller-manager","lastTimestamp":"2025-12-04T11:37:49Z","reason":"FailedScheduling","message":"0/1 nodes are available: 1 node(s) didn't match pod anti-affinity rules. preemption: 0/1 nodes are available: 1 node(s) didn't match pod anti-affinity rules.","type":"Warning"},{"namespace":"openshift-route-controller-manager","lastTimestamp":"2025-12-04T11:37:52Z","reason":"Scheduled","message":"Successfully assigned openshift-route-controller-manager/route-controller-manager-5f845c897b-hhhv6 to master-0","type":"Normal"},{"namespace":"openshift-route-controller-manager","lastTimestamp":"2025-12-04T11:38:00Z","reason":"FailedMount","message":"MountVolume.SetUp failed for volume \"client-ca\" : configmap \"client-ca\" not found","type":"Warning"},{"namespace":"openshift-route-controller-manager","lastTimestamp":"2025-12-04T11:38:00Z","reason":"FailedMount","message":"MountVolume.SetUp failed for volume \"serving-cert\" : secret \"serving-cert\" not found","type":"Warning"},{"namespace":"openshift-route-controller-manager","lastTimestamp":"2025-12-04T11:38:05Z","reason":"ScalingReplicaSet","message":"Scaled down replica set route-controller-manager-5f845c897b to 0 from 1","type":"Normal"},{"namespace":"openshift-route-controller-manager","lastTimestamp":"2025-12-04T11:38:05Z","reason":"SuccessfulDelete","message":"Deleted pod: route-controller-manager-5f845c897b-hhhv6","type":"Normal"},{"namespace":"openshift-route-controller-manager","lastTimestamp":"2025-12-04T11:38:05Z","reason":"SuccessfulCreate","message":"Created pod: route-controller-manager-f4c4cbbd-5bv8h","type":"Normal"},{"namespace":"openshift-route-controller-manager","lastTimestamp":"2025-12-04T11:38:05Z","reason":"FailedScheduling","message":"0/1 nodes are available: 1 node(s) didn't match pod anti-affinity rules. preemption: 0/1 nodes are available: 1 node(s) didn't match pod anti-affinity rules.","type":"Warning"},{"namespace":"openshift-route-controller-manager","lastTimestamp":"2025-12-04T11:38:05Z","reason":"ScalingReplicaSet","message":"Scaled up replica set route-controller-manager-f4c4cbbd to 1 from 0","type":"Normal"},{"namespace":"openshift-route-controller-manager","lastTimestamp":"2025-12-04T11:38:08Z","reason":"FailedMount","message":"MountVolume.SetUp failed for volume \"client-ca\" : object \"openshift-route-controller-manager\"/\"client-ca\" not registered","type":"Warning"},{"namespace":"openshift-route-controller-manager","lastTimestamp":"2025-12-04T11:38:08Z","reason":"FailedMount","message":"MountVolume.SetUp failed for volume \"serving-cert\" : object \"openshift-route-controller-manager\"/\"serving-cert\" not registered","type":"Warning"},{"namespace":"openshift-route-controller-manager","lastTimestamp":"2025-12-04T11:38:12Z","reason":"Scheduled","message":"Successfully assigned openshift-route-controller-manager/route-controller-manager-f4c4cbbd-5bv8h to master-0","type":"Normal"},{"namespace":"openshift-route-controller-manager","lastTimestamp":"2025-12-04T11:38:19Z","reason":"AddedInterface","message":"Add eth0 [10.128.0.37/23] from ovn-kubernetes","type":"Normal"},{"namespace":"openshift-route-controller-manager","lastTimestamp":"2025-12-04T11:38:19Z","reason":"Pulling","message":"Pulling image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c416b201d480bddb5a4960ec42f4740761a1335001cf84ba5ae19ad6857771b1\"","type":"Normal"},{"namespace":"openshift-route-controller-manager","lastTimestamp":"2025-12-04T11:38:24Z","reason":"ScalingReplicaSet","message":"Scaled down replica set route-controller-manager-f4c4cbbd to 0 from 1","type":"Normal"},{"namespace":"openshift-route-controller-manager","lastTimestamp":"2025-12-04T11:38:24Z","reason":"SuccessfulDelete","message":"Deleted pod: route-controller-manager-f4c4cbbd-5bv8h","type":"Normal"},{"namespace":"openshift-route-controller-manager","lastTimestamp":"2025-12-04T11:38:24Z","reason":"ScalingReplicaSet","message":"Scaled up replica set route-controller-manager-846b467b5c to 1 from 0","type":"Normal"},{"namespace":"openshift-route-controller-manager","lastTimestamp":"2025-12-04T11:38:25Z","reason":"SuccessfulCreate","message":"Created pod: route-controller-manager-846b467b5c-thc5v","type":"Normal"},{"namespace":"openshift-route-controller-manager","lastTimestamp":"2025-12-04T11:38:32Z","reason":"Created","message":"Created container: route-controller-manager","type":"Normal"},{"namespace":"openshift-route-controller-manager","lastTimestamp":"2025-12-04T11:38:32Z","reason":"Started","message":"Started container route-controller-manager","type":"Normal"},{"namespace":"openshift-route-controller-manager","lastTimestamp":"2025-12-04T11:38:32Z","reason":"Pulled","message":"Successfully pulled image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c416b201d480bddb5a4960ec42f4740761a1335001cf84ba5ae19ad6857771b1\" in 12.348s (12.348s including waiting). Image size: 481559117 bytes.","type":"Normal"},{"namespace":"openshift-route-controller-manager","lastTimestamp":"2025-12-04T11:38:33Z","reason":"Killing","message":"Stopping container route-controller-manager","type":"Normal"},{"namespace":"openshift-route-controller-manager","lastTimestamp":"2025-12-04T11:38:33Z","reason":"LeaderElection","message":"route-controller-manager-f4c4cbbd-5bv8h_4f9b9748-3e03-4be7-91d8-9b9b5396130d became leader","type":"Normal"},{"namespace":"openshift-route-controller-manager","lastTimestamp":"2025-12-04T11:38:34Z","reason":"FailedScheduling","message":"0/1 nodes are available: 1 node(s) didn't match pod anti-affinity rules. preemption: 0/1 nodes are available: 1 node(s) didn't match pod anti-affinity rules.","type":"Warning"},{"namespace":"openshift-route-controller-manager","lastTimestamp":"2025-12-04T11:38:36Z","reason":"Scheduled","message":"Successfully assigned openshift-route-controller-manager/route-controller-manager-846b467b5c-thc5v to master-0","type":"Normal"},{"namespace":"openshift-route-controller-manager","lastTimestamp":"2025-12-04T11:38:37Z","reason":"AddedInterface","message":"Add eth0 [10.128.0.53/23] from ovn-kubernetes","type":"Normal"},{"namespace":"openshift-route-controller-manager","lastTimestamp":"2025-12-04T11:38:38Z","reason":"Pulled","message":"Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c416b201d480bddb5a4960ec42f4740761a1335001cf84ba5ae19ad6857771b1\" already present on machine","type":"Normal"},{"namespace":"openshift-route-controller-manager","lastTimestamp":"2025-12-04T11:38:40Z","reason":"Started","message":"Started container route-controller-manager","type":"Normal"},{"namespace":"openshift-route-controller-manager","lastTimestamp":"2025-12-04T11:38:40Z","reason":"Created","message":"Created container: route-controller-manager","type":"Normal"},{"namespace":"openshift-route-controller-manager","lastTimestamp":"2025-12-04T11:38:41Z","reason":"LeaderElection","message":"route-controller-manager-846b467b5c-thc5v_e6929345-e307-497a-bdf0-99ff141b1762 became leader","type":"Normal"},{"namespace":"openshift-route-controller-manager","lastTimestamp":"2025-12-04T11:38:45Z","reason":"ScalingReplicaSet","message":"Scaled down replica set route-controller-manager-846b467b5c to 0 from 1","type":"Normal"},{"namespace":"openshift-route-controller-manager","lastTimestamp":"2025-12-04T11:38:45Z","reason":"SuccessfulDelete","message":"Deleted pod: route-controller-manager-846b467b5c-thc5v","type":"Normal"},{"namespace":"openshift-route-controller-manager","lastTimestamp":"2025-12-04T11:38:45Z","reason":"Killing","message":"Stopping container route-controller-manager","type":"Normal"},{"namespace":"openshift-route-controller-manager","lastTimestamp":"2025-12-04T11:38:45Z","reason":"SuccessfulCreate","message":"Created pod: route-controller-manager-95cb5f987-46bsk","type":"Normal"},{"namespace":"openshift-route-controller-manager","lastTimestamp":"2025-12-04T11:38:45Z","reason":"ScalingReplicaSet","message":"Scaled up replica set route-controller-manager-95cb5f987 to 1 from 0","type":"Normal"},{"namespace":"openshift-route-controller-manager","lastTimestamp":"2025-12-04T11:38:48Z","reason":"FailedScheduling","message":"0/1 nodes are available: 1 node(s) didn't match pod anti-affinity rules. preemption: 0/1 nodes are available: 1 node(s) didn't match pod anti-affinity rules.","type":"Warning"},{"namespace":"openshift-route-controller-manager","lastTimestamp":"2025-12-04T11:38:48Z","reason":"Unhealthy","message":"Readiness probe failed: Get \"https://10.128.0.53:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)","type":"Warning"},{"namespace":"openshift-route-controller-manager","lastTimestamp":"2025-12-04T11:38:48Z","reason":"ProbeError","message":"Readiness probe error: Get \"https://10.128.0.53:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)\nbody: \n","type":"Warning"},{"namespace":"openshift-route-controller-manager","lastTimestamp":"2025-12-04T11:38:50Z","reason":"Scheduled","message":"Successfully assigned openshift-route-controller-manager/route-controller-manager-95cb5f987-46bsk to master-0","type":"Normal"},{"namespace":"openshift-route-controller-manager","lastTimestamp":"2025-12-04T11:40:22Z","reason":"FailedCreatePodSandBox","message":"Failed to create pod sandbox: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_route-controller-manager-95cb5f987-46bsk_openshift-route-controller-manager_e4d7939a-5961-4608-b910-73e71aa55bf6_0(9519a074afd2181f00fe2f6116d50a06885facad038c0cb4784a9787e54f04e8): error adding pod openshift-route-controller-manager_route-controller-manager-95cb5f987-46bsk to CNI network \"multus-cni-network\": plugin type=\"multus-shim\" name=\"multus-cni-network\" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:\"9519a074afd2181f00fe2f6116d50a06885facad038c0cb4784a9787e54f04e8\" Netns:\"/var/run/netns/e22c2237-8bd6-44a7-8811-bb69ea806f2b\" IfName:\"eth0\" Args:\"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-route-controller-manager;K8S_POD_NAME=route-controller-manager-95cb5f987-46bsk;K8S_POD_INFRA_CONTAINER_ID=9519a074afd2181f00fe2f6116d50a06885facad038c0cb4784a9787e54f04e8;K8S_POD_UID=e4d7939a-5961-4608-b910-73e71aa55bf6\" Path:\"\" ERRORED: error configuring pod [openshift-route-controller-manager/route-controller-manager-95cb5f987-46bsk] networking: Multus: [openshift-route-controller-manager/route-controller-manager-95cb5f987-46bsk/e4d7939a-5961-4608-b910-73e71aa55bf6]: error setting the networks status: SetPodNetworkStatusAnnotation: failed to update the pod route-controller-manager-95cb5f987-46bsk in out of cluster comm: SetNetworkStatus: failed to update the pod route-controller-manager-95cb5f987-46bsk in out of cluster comm: status update failed for pod /: Get \"https://api-int.sno.openstack.lab:6443/api/v1/namespaces/openshift-route-controller-manager/pods/route-controller-manager-95cb5f987-46bsk?timeout=1m0s\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)\n': StdinData: {\"binDir\":\"/var/lib/cni/bin\",\"clusterNetwork\":\"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf\",\"cniVersion\":\"0.3.1\",\"daemonSocketDir\":\"/run/multus/socket\",\"globalNamespaces\":\"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv\",\"logLevel\":\"verbose\",\"logToStderr\":true,\"name\":\"multus-cni-network\",\"namespaceIsolation\":true,\"type\":\"multus-shim\"}","type":"Warning"},{"namespace":"openshift-route-controller-manager","lastTimestamp":"2025-12-04T11:41:23Z","reason":"FailedCreatePodSandBox","message":"Failed to create pod sandbox: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_route-controller-manager-95cb5f987-46bsk_openshift-route-controller-manager_e4d7939a-5961-4608-b910-73e71aa55bf6_0(aea5fc223c451dccb2eb2f140edba281c9bae377ca308d1bba7b28f3cd4f529d): error adding pod openshift-route-controller-manager_route-controller-manager-95cb5f987-46bsk to CNI network \"multus-cni-network\": plugin type=\"multus-shim\" name=\"multus-cni-network\" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:\"aea5fc223c451dccb2eb2f140edba281c9bae377ca308d1bba7b28f3cd4f529d\" Netns:\"/var/run/netns/13e2d887-db77-4ff8-8664-99e6741ac5b6\" IfName:\"eth0\" Args:\"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-route-controller-manager;K8S_POD_NAME=route-controller-manager-95cb5f987-46bsk;K8S_POD_INFRA_CONTAINER_ID=aea5fc223c451dccb2eb2f140edba281c9bae377ca308d1bba7b28f3cd4f529d;K8S_POD_UID=e4d7939a-5961-4608-b910-73e71aa55bf6\" Path:\"\" ERRORED: error configuring pod [openshift-route-controller-manager/route-controller-manager-95cb5f987-46bsk] networking: Multus: [openshift-route-controller-manager/route-controller-manager-95cb5f987-46bsk/e4d7939a-5961-4608-b910-73e71aa55bf6]: error setting the networks status: SetPodNetworkStatusAnnotation: failed to update the pod route-controller-manager-95cb5f987-46bsk in out of cluster comm: SetNetworkStatus: failed to update the pod route-controller-manager-95cb5f987-46bsk in out of cluster comm: status update failed for pod /: Get \"https://api-int.sno.openstack.lab:6443/api/v1/namespaces/openshift-route-controller-manager/pods/route-controller-manager-95cb5f987-46bsk?timeout=1m0s\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)\n': StdinData: {\"binDir\":\"/var/lib/cni/bin\",\"clusterNetwork\":\"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf\",\"cniVersion\":\"0.3.1\",\"daemonSocketDir\":\"/run/multus/socket\",\"globalNamespaces\":\"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv\",\"logLevel\":\"verbose\",\"logToStderr\":true,\"name\":\"multus-cni-network\",\"namespaceIsolation\":true,\"type\":\"multus-shim\"}","type":"Warning"},{"namespace":"openshift-route-controller-manager","lastTimestamp":"2025-12-04T11:43:12Z","reason":"FailedCreatePodSandBox","message":"Failed to create pod sandbox: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_route-controller-manager-95cb5f987-46bsk_openshift-route-controller-manager_e4d7939a-5961-4608-b910-73e71aa55bf6_0(639b179c776f4ece5a98ff46eeb6b6de6d24a404e0d5f5af53ad6080b597a9d1): error adding pod openshift-route-controller-manager_route-controller-manager-95cb5f987-46bsk to CNI network \"multus-cni-network\": plugin type=\"multus-shim\" name=\"multus-cni-network\" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:\"639b179c776f4ece5a98ff46eeb6b6de6d24a404e0d5f5af53ad6080b597a9d1\" Netns:\"/var/run/netns/e33289fb-8fac-4c47-9f5b-3d6ca06ce72b\" IfName:\"eth0\" Args:\"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-route-controller-manager;K8S_POD_NAME=route-controller-manager-95cb5f987-46bsk;K8S_POD_INFRA_CONTAINER_ID=639b179c776f4ece5a98ff46eeb6b6de6d24a404e0d5f5af53ad6080b597a9d1;K8S_POD_UID=e4d7939a-5961-4608-b910-73e71aa55bf6\" Path:\"\" ERRORED: error configuring pod [openshift-route-controller-manager/route-controller-manager-95cb5f987-46bsk] networking: Multus: [openshift-route-controller-manager/route-controller-manager-95cb5f987-46bsk/e4d7939a-5961-4608-b910-73e71aa55bf6]: error setting the networks status: SetPodNetworkStatusAnnotation: failed to update the pod route-controller-manager-95cb5f987-46bsk in out of cluster comm: SetNetworkStatus: failed to update the pod route-controller-manager-95cb5f987-46bsk in out of cluster comm: status update failed for pod /: Get \"https://api-int.sno.openstack.lab:6443/api/v1/namespaces/openshift-route-controller-manager/pods/route-controller-manager-95cb5f987-46bsk?timeout=1m0s\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)\n': StdinData: {\"binDir\":\"/var/lib/cni/bin\",\"clusterNetwork\":\"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf\",\"cniVersion\":\"0.3.1\",\"daemonSocketDir\":\"/run/multus/socket\",\"globalNamespaces\":\"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv\",\"logLevel\":\"verbose\",\"logToStderr\":true,\"name\":\"multus-cni-network\",\"namespaceIsolation\":true,\"type\":\"multus-shim\"}","type":"Warning"},{"namespace":"openshift-route-controller-manager","lastTimestamp":"2025-12-04T11:44:54Z","reason":"FailedCreatePodSandBox","message":"Failed to create pod sandbox: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_route-controller-manager-95cb5f987-46bsk_openshift-route-controller-manager_e4d7939a-5961-4608-b910-73e71aa55bf6_0(6d3b3f16703c335a06514238952905813ef880930cb5654115a32b53928d3f6c): error adding pod openshift-route-controller-manager_route-controller-manager-95cb5f987-46bsk to CNI network \"multus-cni-network\": plugin type=\"multus-shim\" name=\"multus-cni-network\" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:\"6d3b3f16703c335a06514238952905813ef880930cb5654115a32b53928d3f6c\" Netns:\"/var/run/netns/e47c14cb-63b2-4e70-a12c-95b0aeef97d3\" IfName:\"eth0\" Args:\"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-route-controller-manager;K8S_POD_NAME=route-controller-manager-95cb5f987-46bsk;K8S_POD_INFRA_CONTAINER_ID=6d3b3f16703c335a06514238952905813ef880930cb5654115a32b53928d3f6c;K8S_POD_UID=e4d7939a-5961-4608-b910-73e71aa55bf6\" Path:\"\" ERRORED: error configuring pod [openshift-route-controller-manager/route-controller-manager-95cb5f987-46bsk] networking: Multus: [openshift-route-controller-manager/route-controller-manager-95cb5f987-46bsk/e4d7939a-5961-4608-b910-73e71aa55bf6]: error setting the networks status: SetPodNetworkStatusAnnotation: failed to update the pod route-controller-manager-95cb5f987-46bsk in out of cluster comm: SetNetworkStatus: failed to update the pod route-controller-manager-95cb5f987-46bsk in out of cluster comm: status update failed for pod /: Get \"https://api-int.sno.openstack.lab:6443/api/v1/namespaces/openshift-route-controller-manager/pods/route-controller-manager-95cb5f987-46bsk?timeout=1m0s\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)\n': StdinData: {\"binDir\":\"/var/lib/cni/bin\",\"clusterNetwork\":\"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf\",\"cniVersion\":\"0.3.1\",\"daemonSocketDir\":\"/run/multus/socket\",\"globalNamespaces\":\"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv\",\"logLevel\":\"verbose\",\"logToStderr\":true,\"name\":\"multus-cni-network\",\"namespaceIsolation\":true,\"type\":\"multus-shim\"}","type":"Warning"},{"namespace":"openshift-route-controller-manager","lastTimestamp":"2025-12-04T11:45:11Z","reason":"AddedInterface","message":"Add eth0 [10.128.0.70/23] from ovn-kubernetes","type":"Normal"},{"namespace":"openshift-route-controller-manager","lastTimestamp":"2025-12-04T11:45:14Z","reason":"LeaderElection","message":"route-controller-manager-95cb5f987-46bsk_2d76dab9-2560-4395-8d5d-0349ab87e4e6 became leader","type":"Normal"},{"namespace":"openshift-route-controller-manager","lastTimestamp":"2025-12-04T11:55:47Z","reason":"Created","message":"Created container: route-controller-manager","type":"Normal"},{"namespace":"openshift-route-controller-manager","lastTimestamp":"2025-12-04T11:55:47Z","reason":"Started","message":"Started container route-controller-manager","type":"Normal"},{"namespace":"openshift-route-controller-manager","lastTimestamp":"2025-12-04T11:55:50Z","reason":"Unhealthy","message":"Readiness probe failed: Get \"https://10.128.0.70:8443/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)","type":"Warning"},{"namespace":"openshift-route-controller-manager","lastTimestamp":"2025-12-04T11:55:50Z","reason":"ProbeError","message":"Readiness probe error: Get \"https://10.128.0.70:8443/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)\nbody: \n","type":"Warning"},{"namespace":"openshift-route-controller-manager","lastTimestamp":"2025-12-04T11:56:22Z","reason":"Unhealthy","message":"Liveness probe failed: Get \"https://10.128.0.70:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)","type":"Warning"},{"namespace":"openshift-route-controller-manager","lastTimestamp":"2025-12-04T11:56:22Z","reason":"ProbeError","message":"Liveness probe error: Get \"https://10.128.0.70:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)\nbody: \n","type":"Warning"},{"namespace":"openshift-route-controller-manager","lastTimestamp":"2025-12-04T11:56:22Z","reason":"Unhealthy","message":"Readiness probe failed: Get \"https://10.128.0.70:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)","type":"Warning"},{"namespace":"openshift-route-controller-manager","lastTimestamp":"2025-12-04T11:56:22Z","reason":"ProbeError","message":"Readiness probe error: Get \"https://10.128.0.70:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)\nbody: \n","type":"Warning"},{"namespace":"openshift-route-controller-manager","lastTimestamp":"2025-12-04T11:56:31Z","reason":"BackOff","message":"Back-off restarting failed container route-controller-manager in pod route-controller-manager-95cb5f987-46bsk_openshift-route-controller-manager(e4d7939a-5961-4608-b910-73e71aa55bf6)","type":"Warning"},{"namespace":"openshift-route-controller-manager","lastTimestamp":"2025-12-04T11:56:42Z","reason":"Pulled","message":"Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c416b201d480bddb5a4960ec42f4740761a1335001cf84ba5ae19ad6857771b1\" already present on machine","type":"Normal"},{"namespace":"openshift-route-controller-manager","lastTimestamp":"2025-12-04T11:56:43Z","reason":"LeaderElection","message":"route-controller-manager-95cb5f987-46bsk_d647a0da-6f37-43a4-80dd-2a0a58f4cd2c became leader","type":"Normal"},{"namespace":"openshift-route-controller-manager","lastTimestamp":"2025-12-04T11:59:27Z","reason":"FailedMount","message":"MountVolume.SetUp failed for volume \"config\" : failed to sync configmap cache: timed out waiting for the condition","type":"Warning"},{"namespace":"openshift-route-controller-manager","lastTimestamp":"2025-12-04T11:59:27Z","reason":"FailedMount","message":"MountVolume.SetUp failed for volume \"serving-cert\" : failed to sync secret cache: timed out waiting for the condition","type":"Warning"},{"namespace":"openshift-route-controller-manager","lastTimestamp":"2025-12-04T11:59:27Z","reason":"FailedMount","message":"MountVolume.SetUp failed for volume \"client-ca\" : failed to sync configmap cache: timed out waiting for the condition","type":"Warning"},{"namespace":"openshift-route-controller-manager","lastTimestamp":"2025-12-04T12:00:58Z","reason":"Pulled","message":"Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c416b201d480bddb5a4960ec42f4740761a1335001cf84ba5ae19ad6857771b1\" already present on machine","type":"Normal"},{"namespace":"openshift-route-controller-manager","lastTimestamp":"2025-12-04T12:00:59Z","reason":"Started","message":"Started container route-controller-manager","type":"Normal"},{"namespace":"openshift-route-controller-manager","lastTimestamp":"2025-12-04T12:00:59Z","reason":"Created","message":"Created container: route-controller-manager","type":"Normal"},{"namespace":"openshift-route-controller-manager","lastTimestamp":"2025-12-04T12:00:59Z","reason":"LeaderElection","message":"route-controller-manager-95cb5f987-46bsk_397195f7-0562-4c91-89d0-2f6233218cb1 became leader","type":"Normal"}]}events/openshift-controller-manager.json0000640000000000000000000005133515114274105017236 0ustar0000000000000000{"items":[{"namespace":"openshift-controller-manager","lastTimestamp":"2025-12-04T11:37:45Z","reason":"ScalingReplicaSet","message":"Scaled up replica set controller-manager-59948648c9 to 1","type":"Normal"},{"namespace":"openshift-controller-manager","lastTimestamp":"2025-12-04T11:37:45Z","reason":"FailedCreate","message":"Error creating: pods \"controller-manager-59948648c9-\" is forbidden: error looking up service account openshift-controller-manager/openshift-controller-manager-sa: serviceaccount \"openshift-controller-manager-sa\" not found","type":"Warning"},{"namespace":"openshift-controller-manager","lastTimestamp":"2025-12-04T11:37:46Z","reason":"SuccessfulCreate","message":"Created pod: controller-manager-59948648c9-krt4p","type":"Normal"},{"namespace":"openshift-controller-manager","lastTimestamp":"2025-12-04T11:37:46Z","reason":"Scheduled","message":"Successfully assigned openshift-controller-manager/controller-manager-59948648c9-krt4p to master-0","type":"Normal"},{"namespace":"openshift-controller-manager","lastTimestamp":"2025-12-04T11:37:47Z","reason":"FailedMount","message":"MountVolume.SetUp failed for volume \"config\" : configmap \"config\" not found","type":"Warning"},{"namespace":"openshift-controller-manager","lastTimestamp":"2025-12-04T11:37:48Z","reason":"SuccessfulDelete","message":"Deleted pod: controller-manager-59948648c9-krt4p","type":"Normal"},{"namespace":"openshift-controller-manager","lastTimestamp":"2025-12-04T11:37:48Z","reason":"ScalingReplicaSet","message":"Scaled up replica set controller-manager-5c6c4578c to 1 from 0","type":"Normal"},{"namespace":"openshift-controller-manager","lastTimestamp":"2025-12-04T11:37:48Z","reason":"FailedScheduling","message":"0/1 nodes are available: 1 node(s) didn't match pod anti-affinity rules. preemption: 0/1 nodes are available: 1 node(s) didn't match pod anti-affinity rules.","type":"Warning"},{"namespace":"openshift-controller-manager","lastTimestamp":"2025-12-04T11:37:48Z","reason":"SuccessfulCreate","message":"Created pod: controller-manager-5c6c4578c-plvql","type":"Normal"},{"namespace":"openshift-controller-manager","lastTimestamp":"2025-12-04T11:37:48Z","reason":"FailedMount","message":"MountVolume.SetUp failed for volume \"serving-cert\" : secret \"serving-cert\" not found","type":"Warning"},{"namespace":"openshift-controller-manager","lastTimestamp":"2025-12-04T11:37:48Z","reason":"FailedMount","message":"MountVolume.SetUp failed for volume \"client-ca\" : configmap \"client-ca\" not found","type":"Warning"},{"namespace":"openshift-controller-manager","lastTimestamp":"2025-12-04T11:37:48Z","reason":"ScalingReplicaSet","message":"Scaled down replica set controller-manager-59948648c9 to 0 from 1","type":"Normal"},{"namespace":"openshift-controller-manager","lastTimestamp":"2025-12-04T11:37:49Z","reason":"Scheduled","message":"Successfully assigned openshift-controller-manager/controller-manager-5c6c4578c-plvql to master-0","type":"Normal"},{"namespace":"openshift-controller-manager","lastTimestamp":"2025-12-04T11:37:53Z","reason":"FailedMount","message":"MountVolume.SetUp failed for volume \"serving-cert\" : secret \"serving-cert\" not found","type":"Warning"},{"namespace":"openshift-controller-manager","lastTimestamp":"2025-12-04T11:37:57Z","reason":"FailedMount","message":"MountVolume.SetUp failed for volume \"client-ca\" : configmap \"client-ca\" not found","type":"Warning"},{"namespace":"openshift-controller-manager","lastTimestamp":"2025-12-04T11:38:05Z","reason":"ScalingReplicaSet","message":"Scaled up replica set controller-manager-666dcff694 to 1 from 0","type":"Normal"},{"namespace":"openshift-controller-manager","lastTimestamp":"2025-12-04T11:38:05Z","reason":"SuccessfulCreate","message":"Created pod: controller-manager-666dcff694-54zwc","type":"Normal"},{"namespace":"openshift-controller-manager","lastTimestamp":"2025-12-04T11:38:05Z","reason":"SuccessfulDelete","message":"Deleted pod: controller-manager-5c6c4578c-plvql","type":"Normal"},{"namespace":"openshift-controller-manager","lastTimestamp":"2025-12-04T11:38:05Z","reason":"ScalingReplicaSet","message":"Scaled down replica set controller-manager-5c6c4578c to 0 from 1","type":"Normal"},{"namespace":"openshift-controller-manager","lastTimestamp":"2025-12-04T11:38:11Z","reason":"FailedScheduling","message":"0/1 nodes are available: 1 node(s) didn't match pod anti-affinity rules. preemption: 0/1 nodes are available: 1 node(s) didn't match pod anti-affinity rules.","type":"Warning"},{"namespace":"openshift-controller-manager","lastTimestamp":"2025-12-04T11:38:18Z","reason":"Scheduled","message":"Successfully assigned openshift-controller-manager/controller-manager-666dcff694-54zwc to master-0","type":"Normal"},{"namespace":"openshift-controller-manager","lastTimestamp":"2025-12-04T11:38:20Z","reason":"AddedInterface","message":"Add eth0 [10.128.0.38/23] from ovn-kubernetes","type":"Normal"},{"namespace":"openshift-controller-manager","lastTimestamp":"2025-12-04T11:38:20Z","reason":"Pulling","message":"Pulling image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:eddedae7578d79b5a3f748000ae5c00b9f14a04710f9f9ec7b52fc569be5dfb8\"","type":"Normal"},{"namespace":"openshift-controller-manager","lastTimestamp":"2025-12-04T11:38:24Z","reason":"FailedScheduling","message":"0/1 nodes are available: 1 node(s) didn't match pod anti-affinity rules. preemption: 0/1 nodes are available: 1 node(s) didn't match pod anti-affinity rules.","type":"Warning"},{"namespace":"openshift-controller-manager","lastTimestamp":"2025-12-04T11:38:24Z","reason":"SuccessfulDelete","message":"Deleted pod: controller-manager-666dcff694-54zwc","type":"Normal"},{"namespace":"openshift-controller-manager","lastTimestamp":"2025-12-04T11:38:24Z","reason":"SuccessfulCreate","message":"Created pod: controller-manager-7d958ff6f6-b8lzt","type":"Normal"},{"namespace":"openshift-controller-manager","lastTimestamp":"2025-12-04T11:38:24Z","reason":"ScalingReplicaSet","message":"Scaled up replica set controller-manager-7d958ff6f6 to 1 from 0","type":"Normal"},{"namespace":"openshift-controller-manager","lastTimestamp":"2025-12-04T11:38:24Z","reason":"ScalingReplicaSet","message":"Scaled down replica set controller-manager-666dcff694 to 0 from 1","type":"Normal"},{"namespace":"openshift-controller-manager","lastTimestamp":"2025-12-04T11:38:30Z","reason":"Pulled","message":"Successfully pulled image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:eddedae7578d79b5a3f748000ae5c00b9f14a04710f9f9ec7b52fc569be5dfb8\" in 10.188s (10.188s including waiting). Image size: 552673986 bytes.","type":"Normal"},{"namespace":"openshift-controller-manager","lastTimestamp":"2025-12-04T11:38:32Z","reason":"LeaderElection","message":"controller-manager-666dcff694-54zwc became leader","type":"Normal"},{"namespace":"openshift-controller-manager","lastTimestamp":"2025-12-04T11:38:32Z","reason":"Created","message":"Created container: controller-manager","type":"Normal"},{"namespace":"openshift-controller-manager","lastTimestamp":"2025-12-04T11:38:32Z","reason":"Started","message":"Started container controller-manager","type":"Normal"},{"namespace":"openshift-controller-manager","lastTimestamp":"2025-12-04T11:38:33Z","reason":"Killing","message":"Stopping container controller-manager","type":"Normal"},{"namespace":"openshift-controller-manager","lastTimestamp":"2025-12-04T11:38:34Z","reason":"Scheduled","message":"Successfully assigned openshift-controller-manager/controller-manager-7d958ff6f6-b8lzt to master-0","type":"Normal"},{"namespace":"openshift-controller-manager","lastTimestamp":"2025-12-04T11:38:35Z","reason":"LeaderElection","message":"controller-manager-7d958ff6f6-b8lzt became leader","type":"Normal"},{"namespace":"openshift-controller-manager","lastTimestamp":"2025-12-04T11:38:35Z","reason":"Started","message":"Started container controller-manager","type":"Normal"},{"namespace":"openshift-controller-manager","lastTimestamp":"2025-12-04T11:38:35Z","reason":"Created","message":"Created container: controller-manager","type":"Normal"},{"namespace":"openshift-controller-manager","lastTimestamp":"2025-12-04T11:38:35Z","reason":"Pulled","message":"Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:eddedae7578d79b5a3f748000ae5c00b9f14a04710f9f9ec7b52fc569be5dfb8\" already present on machine","type":"Normal"},{"namespace":"openshift-controller-manager","lastTimestamp":"2025-12-04T11:38:35Z","reason":"AddedInterface","message":"Add eth0 [10.128.0.50/23] from ovn-kubernetes","type":"Normal"},{"namespace":"openshift-controller-manager","lastTimestamp":"2025-12-04T11:38:45Z","reason":"Killing","message":"Stopping container controller-manager","type":"Normal"},{"namespace":"openshift-controller-manager","lastTimestamp":"2025-12-04T11:38:45Z","reason":"SuccessfulCreate","message":"Created pod: controller-manager-6686654b8d-rrndk","type":"Normal"},{"namespace":"openshift-controller-manager","lastTimestamp":"2025-12-04T11:38:45Z","reason":"FailedScheduling","message":"0/1 nodes are available: 1 node(s) didn't match pod anti-affinity rules. preemption: 0/1 nodes are available: 1 node(s) didn't match pod anti-affinity rules.","type":"Warning"},{"namespace":"openshift-controller-manager","lastTimestamp":"2025-12-04T11:38:45Z","reason":"ScalingReplicaSet","message":"Scaled up replica set controller-manager-6686654b8d to 1 from 0","type":"Normal"},{"namespace":"openshift-controller-manager","lastTimestamp":"2025-12-04T11:38:45Z","reason":"ScalingReplicaSet","message":"Scaled down replica set controller-manager-7d958ff6f6 to 0 from 1","type":"Normal"},{"namespace":"openshift-controller-manager","lastTimestamp":"2025-12-04T11:38:45Z","reason":"SuccessfulDelete","message":"Deleted pod: controller-manager-7d958ff6f6-b8lzt","type":"Normal"},{"namespace":"openshift-controller-manager","lastTimestamp":"2025-12-04T11:38:48Z","reason":"Scheduled","message":"Successfully assigned openshift-controller-manager/controller-manager-6686654b8d-rrndk to master-0","type":"Normal"},{"namespace":"openshift-controller-manager","lastTimestamp":"2025-12-04T11:40:22Z","reason":"FailedCreatePodSandBox","message":"Failed to create pod sandbox: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_controller-manager-6686654b8d-rrndk_openshift-controller-manager_24506aa4-ab78-49df-bb58-59093498f13d_0(80084e6d846b9ab14d689920224de665b033fb5b2fda92ba002f913e4ec488c5): error adding pod openshift-controller-manager_controller-manager-6686654b8d-rrndk to CNI network \"multus-cni-network\": plugin type=\"multus-shim\" name=\"multus-cni-network\" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:\"80084e6d846b9ab14d689920224de665b033fb5b2fda92ba002f913e4ec488c5\" Netns:\"/var/run/netns/d7e501b9-d461-4a07-adb8-dcf7d33c4eaf\" IfName:\"eth0\" Args:\"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-controller-manager;K8S_POD_NAME=controller-manager-6686654b8d-rrndk;K8S_POD_INFRA_CONTAINER_ID=80084e6d846b9ab14d689920224de665b033fb5b2fda92ba002f913e4ec488c5;K8S_POD_UID=24506aa4-ab78-49df-bb58-59093498f13d\" Path:\"\" ERRORED: error configuring pod [openshift-controller-manager/controller-manager-6686654b8d-rrndk] networking: Multus: [openshift-controller-manager/controller-manager-6686654b8d-rrndk/24506aa4-ab78-49df-bb58-59093498f13d]: error setting the networks status: SetPodNetworkStatusAnnotation: failed to update the pod controller-manager-6686654b8d-rrndk in out of cluster comm: SetNetworkStatus: failed to update the pod controller-manager-6686654b8d-rrndk in out of cluster comm: status update failed for pod /: Get \"https://api-int.sno.openstack.lab:6443/api/v1/namespaces/openshift-controller-manager/pods/controller-manager-6686654b8d-rrndk?timeout=1m0s\": context deadline exceeded\n': StdinData: {\"binDir\":\"/var/lib/cni/bin\",\"clusterNetwork\":\"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf\",\"cniVersion\":\"0.3.1\",\"daemonSocketDir\":\"/run/multus/socket\",\"globalNamespaces\":\"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv\",\"logLevel\":\"verbose\",\"logToStderr\":true,\"name\":\"multus-cni-network\",\"namespaceIsolation\":true,\"type\":\"multus-shim\"}","type":"Warning"},{"namespace":"openshift-controller-manager","lastTimestamp":"2025-12-04T11:41:23Z","reason":"FailedCreatePodSandBox","message":"Failed to create pod sandbox: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_controller-manager-6686654b8d-rrndk_openshift-controller-manager_24506aa4-ab78-49df-bb58-59093498f13d_0(dd455792c3c0c609a825558a90a6fdaff70c7f5c28eaeed93416f3bf654b7a11): error adding pod openshift-controller-manager_controller-manager-6686654b8d-rrndk to CNI network \"multus-cni-network\": plugin type=\"multus-shim\" name=\"multus-cni-network\" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:\"dd455792c3c0c609a825558a90a6fdaff70c7f5c28eaeed93416f3bf654b7a11\" Netns:\"/var/run/netns/0314a257-ead8-4ab3-afce-07d51d75d96d\" IfName:\"eth0\" Args:\"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-controller-manager;K8S_POD_NAME=controller-manager-6686654b8d-rrndk;K8S_POD_INFRA_CONTAINER_ID=dd455792c3c0c609a825558a90a6fdaff70c7f5c28eaeed93416f3bf654b7a11;K8S_POD_UID=24506aa4-ab78-49df-bb58-59093498f13d\" Path:\"\" ERRORED: error configuring pod [openshift-controller-manager/controller-manager-6686654b8d-rrndk] networking: Multus: [openshift-controller-manager/controller-manager-6686654b8d-rrndk/24506aa4-ab78-49df-bb58-59093498f13d]: error setting the networks status: SetPodNetworkStatusAnnotation: failed to update the pod controller-manager-6686654b8d-rrndk in out of cluster comm: SetNetworkStatus: failed to update the pod controller-manager-6686654b8d-rrndk in out of cluster comm: status update failed for pod /: Get \"https://api-int.sno.openstack.lab:6443/api/v1/namespaces/openshift-controller-manager/pods/controller-manager-6686654b8d-rrndk?timeout=1m0s\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)\n': StdinData: {\"binDir\":\"/var/lib/cni/bin\",\"clusterNetwork\":\"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf\",\"cniVersion\":\"0.3.1\",\"daemonSocketDir\":\"/run/multus/socket\",\"globalNamespaces\":\"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv\",\"logLevel\":\"verbose\",\"logToStderr\":true,\"name\":\"multus-cni-network\",\"namespaceIsolation\":true,\"type\":\"multus-shim\"}","type":"Warning"},{"namespace":"openshift-controller-manager","lastTimestamp":"2025-12-04T11:42:38Z","reason":"FailedCreatePodSandBox","message":"Failed to create pod sandbox: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_controller-manager-6686654b8d-rrndk_openshift-controller-manager_24506aa4-ab78-49df-bb58-59093498f13d_0(2d160b6d52d8b992b0177da71ff6a4532495aa60365c9fbd31af5eef7c1b5925): error adding pod openshift-controller-manager_controller-manager-6686654b8d-rrndk to CNI network \"multus-cni-network\": plugin type=\"multus-shim\" name=\"multus-cni-network\" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:\"2d160b6d52d8b992b0177da71ff6a4532495aa60365c9fbd31af5eef7c1b5925\" Netns:\"/var/run/netns/e76f4cd1-de22-49cc-8cce-dc93a58a95e5\" IfName:\"eth0\" Args:\"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-controller-manager;K8S_POD_NAME=controller-manager-6686654b8d-rrndk;K8S_POD_INFRA_CONTAINER_ID=2d160b6d52d8b992b0177da71ff6a4532495aa60365c9fbd31af5eef7c1b5925;K8S_POD_UID=24506aa4-ab78-49df-bb58-59093498f13d\" Path:\"\" ERRORED: error configuring pod [openshift-controller-manager/controller-manager-6686654b8d-rrndk] networking: Multus: [openshift-controller-manager/controller-manager-6686654b8d-rrndk/24506aa4-ab78-49df-bb58-59093498f13d]: error setting the networks status: SetPodNetworkStatusAnnotation: failed to update the pod controller-manager-6686654b8d-rrndk in out of cluster comm: SetNetworkStatus: failed to update the pod controller-manager-6686654b8d-rrndk in out of cluster comm: status update failed for pod /: Get \"https://api-int.sno.openstack.lab:6443/api/v1/namespaces/openshift-controller-manager/pods/controller-manager-6686654b8d-rrndk?timeout=1m0s\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)\n': StdinData: {\"binDir\":\"/var/lib/cni/bin\",\"clusterNetwork\":\"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf\",\"cniVersion\":\"0.3.1\",\"daemonSocketDir\":\"/run/multus/socket\",\"globalNamespaces\":\"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv\",\"logLevel\":\"verbose\",\"logToStderr\":true,\"name\":\"multus-cni-network\",\"namespaceIsolation\":true,\"type\":\"multus-shim\"}","type":"Warning"},{"namespace":"openshift-controller-manager","lastTimestamp":"2025-12-04T11:44:20Z","reason":"FailedCreatePodSandBox","message":"Failed to create pod sandbox: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_controller-manager-6686654b8d-rrndk_openshift-controller-manager_24506aa4-ab78-49df-bb58-59093498f13d_0(cfd724b83db7eda66758ee8447129a7dfeb31809a751bc0358c6f09a9428e9ee): error adding pod openshift-controller-manager_controller-manager-6686654b8d-rrndk to CNI network \"multus-cni-network\": plugin type=\"multus-shim\" name=\"multus-cni-network\" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:\"cfd724b83db7eda66758ee8447129a7dfeb31809a751bc0358c6f09a9428e9ee\" Netns:\"/var/run/netns/189d0a6b-f161-4d53-a497-b9eb33e6c80e\" IfName:\"eth0\" Args:\"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-controller-manager;K8S_POD_NAME=controller-manager-6686654b8d-rrndk;K8S_POD_INFRA_CONTAINER_ID=cfd724b83db7eda66758ee8447129a7dfeb31809a751bc0358c6f09a9428e9ee;K8S_POD_UID=24506aa4-ab78-49df-bb58-59093498f13d\" Path:\"\" ERRORED: error configuring pod [openshift-controller-manager/controller-manager-6686654b8d-rrndk] networking: Multus: [openshift-controller-manager/controller-manager-6686654b8d-rrndk/24506aa4-ab78-49df-bb58-59093498f13d]: error setting the networks status: SetPodNetworkStatusAnnotation: failed to update the pod controller-manager-6686654b8d-rrndk in out of cluster comm: SetNetworkStatus: failed to update the pod controller-manager-6686654b8d-rrndk in out of cluster comm: status update failed for pod /: Get \"https://api-int.sno.openstack.lab:6443/api/v1/namespaces/openshift-controller-manager/pods/controller-manager-6686654b8d-rrndk?timeout=1m0s\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)\n': StdinData: {\"binDir\":\"/var/lib/cni/bin\",\"clusterNetwork\":\"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf\",\"cniVersion\":\"0.3.1\",\"daemonSocketDir\":\"/run/multus/socket\",\"globalNamespaces\":\"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv\",\"logLevel\":\"verbose\",\"logToStderr\":true,\"name\":\"multus-cni-network\",\"namespaceIsolation\":true,\"type\":\"multus-shim\"}","type":"Warning"},{"namespace":"openshift-controller-manager","lastTimestamp":"2025-12-04T11:45:02Z","reason":"AddedInterface","message":"Add eth0 [10.128.0.69/23] from ovn-kubernetes","type":"Normal"},{"namespace":"openshift-controller-manager","lastTimestamp":"2025-12-04T11:45:11Z","reason":"LeaderElection","message":"controller-manager-6686654b8d-rrndk became leader","type":"Normal"},{"namespace":"openshift-controller-manager","lastTimestamp":"2025-12-04T11:53:38Z","reason":"Pulled","message":"Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:eddedae7578d79b5a3f748000ae5c00b9f14a04710f9f9ec7b52fc569be5dfb8\" already present on machine","type":"Normal"},{"namespace":"openshift-controller-manager","lastTimestamp":"2025-12-04T11:53:38Z","reason":"Created","message":"Created container: controller-manager","type":"Normal"},{"namespace":"openshift-controller-manager","lastTimestamp":"2025-12-04T11:53:38Z","reason":"Started","message":"Started container controller-manager","type":"Normal"},{"namespace":"openshift-controller-manager","lastTimestamp":"2025-12-04T11:55:46Z","reason":"LeaderElection","message":"controller-manager-6686654b8d-rrndk became leader","type":"Normal"},{"namespace":"openshift-controller-manager","lastTimestamp":"2025-12-04T11:59:27Z","reason":"FailedMount","message":"MountVolume.SetUp failed for volume \"client-ca\" : failed to sync configmap cache: timed out waiting for the condition","type":"Warning"},{"namespace":"openshift-controller-manager","lastTimestamp":"2025-12-04T11:59:27Z","reason":"FailedMount","message":"MountVolume.SetUp failed for volume \"proxy-ca-bundles\" : failed to sync configmap cache: timed out waiting for the condition","type":"Warning"},{"namespace":"openshift-controller-manager","lastTimestamp":"2025-12-04T11:59:27Z","reason":"FailedMount","message":"MountVolume.SetUp failed for volume \"serving-cert\" : failed to sync secret cache: timed out waiting for the condition","type":"Warning"},{"namespace":"openshift-controller-manager","lastTimestamp":"2025-12-04T11:59:27Z","reason":"FailedMount","message":"MountVolume.SetUp failed for volume \"config\" : failed to sync configmap cache: timed out waiting for the condition","type":"Warning"}]}events/openshift-controller-manager-operator.json0000640000000000000000000012770515114274105021074 0ustar0000000000000000{"items":[{"namespace":"openshift-controller-manager-operator","lastTimestamp":"2025-12-04T11:31:30Z","reason":"ScalingReplicaSet","message":"Scaled up replica set openshift-controller-manager-operator-6c8676f99d to 1","type":"Normal"},{"namespace":"openshift-controller-manager-operator","lastTimestamp":"2025-12-04T11:31:41Z","reason":"FailedCreate","message":"Error creating: pods \"openshift-controller-manager-operator-6c8676f99d-\" is forbidden: autoscaling.openshift.io/ManagementCPUsOverride the cluster does not have any nodes","type":"Warning"},{"namespace":"openshift-controller-manager-operator","lastTimestamp":"2025-12-04T11:35:47Z","reason":"FailedCreate","message":"Error creating: pods \"openshift-controller-manager-operator-6c8676f99d-\" is forbidden: autoscaling.openshift.io/ManagementCPUsOverride the cluster does not have any nodes","type":"Warning"},{"namespace":"openshift-controller-manager-operator","lastTimestamp":"2025-12-04T11:35:49Z","reason":"SuccessfulCreate","message":"Created pod: openshift-controller-manager-operator-6c8676f99d-7z948","type":"Normal"},{"namespace":"openshift-controller-manager-operator","lastTimestamp":"2025-12-04T11:35:49Z","reason":"FailedScheduling","message":"0/1 nodes are available: 1 node(s) had untolerated taint {node.kubernetes.io/not-ready: }. preemption: 0/1 nodes are available: 1 Preemption is not helpful for scheduling.","type":"Warning"},{"namespace":"openshift-controller-manager-operator","lastTimestamp":"2025-12-04T11:37:30Z","reason":"Scheduled","message":"Successfully assigned openshift-controller-manager-operator/openshift-controller-manager-operator-6c8676f99d-7z948 to master-0","type":"Normal"},{"namespace":"openshift-controller-manager-operator","lastTimestamp":"2025-12-04T11:37:33Z","reason":"AddedInterface","message":"Add eth0 [10.128.0.8/23] from ovn-kubernetes","type":"Normal"},{"namespace":"openshift-controller-manager-operator","lastTimestamp":"2025-12-04T11:37:33Z","reason":"Pulling","message":"Pulling image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8eabac819f289e29d75c7ab172d8124554849a47f0b00770928c3eb19a5a31c4\"","type":"Normal"},{"namespace":"openshift-controller-manager-operator","lastTimestamp":"2025-12-04T11:37:42Z","reason":"Started","message":"Started container openshift-controller-manager-operator","type":"Normal"},{"namespace":"openshift-controller-manager-operator","lastTimestamp":"2025-12-04T11:37:42Z","reason":"Created","message":"Created container: openshift-controller-manager-operator","type":"Normal"},{"namespace":"openshift-controller-manager-operator","lastTimestamp":"2025-12-04T11:37:42Z","reason":"Pulled","message":"Successfully pulled image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8eabac819f289e29d75c7ab172d8124554849a47f0b00770928c3eb19a5a31c4\" in 8.757s (8.757s including waiting). Image size: 502436444 bytes.","type":"Normal"},{"namespace":"openshift-controller-manager-operator","lastTimestamp":"2025-12-04T11:37:44Z","reason":"ObservedConfigChanged","message":"Writing updated observed config: map[string]any{\n+ \t\"build\": map[string]any{\n+ \t\t\"buildDefaults\": map[string]any{\"resources\": map[string]any{}},\n+ \t\t\"imageTemplateFormat\": map[string]any{\n+ \t\t\t\"format\": string(\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:31aa3c7464\"...),\n+ \t\t},\n+ \t},\n+ \t\"controllers\": []any{\n+ \t\tstring(\"openshift.io/build\"), string(\"openshift.io/build-config-change\"),\n+ \t\tstring(\"openshift.io/builder-rolebindings\"),\n+ \t\tstring(\"openshift.io/builder-serviceaccount\"),\n+ \t\tstring(\"-openshift.io/default-rolebindings\"), string(\"openshift.io/deployer\"),\n+ \t\tstring(\"openshift.io/deployer-rolebindings\"),\n+ \t\tstring(\"openshift.io/deployer-serviceaccount\"),\n+ \t\tstring(\"openshift.io/deploymentconfig\"), string(\"openshift.io/image-import\"),\n+ \t\tstring(\"openshift.io/image-puller-rolebindings\"),\n+ \t\tstring(\"openshift.io/image-signature-import\"),\n+ \t\tstring(\"openshift.io/image-trigger\"), string(\"openshift.io/ingress-ip\"),\n+ \t\tstring(\"openshift.io/ingress-to-route\"),\n+ \t\tstring(\"openshift.io/origin-namespace\"), ...,\n+ \t},\n+ \t\"deployer\": map[string]any{\n+ \t\t\"imageTemplateFormat\": map[string]any{\n+ \t\t\t\"format\": string(\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:42c3f5030d\"...),\n+ \t\t},\n+ \t},\n+ \t\"featureGates\": []any{string(\"BuildCSIVolumes=true\")},\n+ \t\"ingress\": map[string]any{\"ingressIPNetworkCIDR\": string(\"\")},\n }\n","type":"Normal"},{"namespace":"openshift-controller-manager-operator","lastTimestamp":"2025-12-04T11:37:44Z","reason":"LeaderElection","message":"openshift-controller-manager-operator-6c8676f99d-7z948_f13ff4b0-c7ac-42bd-ab3b-8b2d428a1859 became leader","type":"Normal"},{"namespace":"openshift-controller-manager-operator","lastTimestamp":"2025-12-04T11:37:44Z","reason":"ClusterRoleCreated","message":"Created ClusterRole.rbac.authorization.k8s.io/system:openshift:tokenreview-openshift-controller-manager because it was missing","type":"Normal"},{"namespace":"openshift-controller-manager-operator","lastTimestamp":"2025-12-04T11:37:44Z","reason":"ClusterRoleBindingCreated","message":"Created ClusterRoleBinding.rbac.authorization.k8s.io/system:openshift:openshift-controller-manager because it was missing","type":"Normal"},{"namespace":"openshift-controller-manager-operator","lastTimestamp":"2025-12-04T11:37:44Z","reason":"OperatorStatusChanged","message":"Status for clusteroperator/openshift-controller-manager changed: Degraded set to Unknown (\"\"),Progressing set to Unknown (\"\"),Available set to Unknown (\"\"),Upgradeable set to Unknown (\"\"),EvaluationConditionsDetected set to Unknown (\"\"),status.relatedObjects changed from [] to [{\"operator.openshift.io\" \"openshiftcontrollermanagers\" \"\" \"cluster\"} {\"\" \"namespaces\" \"\" \"openshift-config\"} {\"\" \"namespaces\" \"\" \"openshift-config-managed\"} {\"\" \"namespaces\" \"\" \"openshift-controller-manager-operator\"} {\"\" \"namespaces\" \"\" \"openshift-controller-manager\"} {\"\" \"namespaces\" \"\" \"openshift-route-controller-manager\"}]","type":"Normal"},{"namespace":"openshift-controller-manager-operator","lastTimestamp":"2025-12-04T11:37:44Z","reason":"ClusterRoleCreated","message":"Created ClusterRole.rbac.authorization.k8s.io/system:openshift:openshift-controller-manager because it was missing","type":"Normal"},{"namespace":"openshift-controller-manager-operator","lastTimestamp":"2025-12-04T11:37:44Z","reason":"FeatureGatesInitialized","message":"FeatureGates updated to featuregates.Features{Enabled:[]v1.FeatureGateName{\"AWSEFSDriverVolumeMetrics\", \"AdminNetworkPolicy\", \"AlibabaPlatform\", \"AzureWorkloadIdentity\", \"BareMetalLoadBalancer\", \"BuildCSIVolumes\", \"ChunkSizeMiB\", \"CloudDualStackNodeIPs\", \"DisableKubeletCloudCredentialProviders\", \"GCPLabelsTags\", \"HardwareSpeed\", \"IngressControllerLBSubnetsAWS\", \"KMSv1\", \"ManagedBootImages\", \"ManagedBootImagesAWS\", \"MultiArchInstallAWS\", \"MultiArchInstallGCP\", \"NetworkDiagnosticsConfig\", \"NetworkLiveMigration\", \"NetworkSegmentation\", \"NewOLM\", \"NodeDisruptionPolicy\", \"OnClusterBuild\", \"PersistentIPsForVirtualization\", \"PrivateHostedZoneAWS\", \"SetEIPForNLBIngressController\", \"VSphereControlPlaneMachineSet\", \"VSphereDriverConfiguration\", \"VSphereMultiVCenters\", \"VSphereStaticIPs\", \"ValidatingAdmissionPolicy\"}, Disabled:[]v1.FeatureGateName{\"AWSClusterHostedDNS\", \"AdditionalRoutingCapabilities\", \"AutomatedEtcdBackup\", \"BootcNodeManagement\", \"CSIDriverSharedResource\", \"ClusterAPIInstall\", \"ClusterAPIInstallIBMCloud\", \"ClusterMonitoringConfig\", \"ConsolePluginContentSecurityPolicy\", \"DNSNameResolver\", \"DynamicResourceAllocation\", \"EtcdBackendQuota\", \"EventedPLEG\", \"Example\", \"ExternalOIDC\", \"ExternalOIDCWithUIDAndExtraClaimMappings\", \"GCPClusterHostedDNS\", \"GatewayAPI\", \"ImageStreamImportMode\", \"IngressControllerDynamicConfigurationManager\", \"InsightsConfig\", \"InsightsConfigAPI\", \"InsightsOnDemandDataGather\", \"InsightsRuntimeExtractor\", \"MachineAPIMigration\", \"MachineAPIOperatorDisableMachineHealthCheckController\", \"MachineAPIProviderOpenStack\", \"MachineConfigNodes\", \"MaxUnavailableStatefulSet\", \"MetricsCollectionProfiles\", \"MinimumKubeletVersion\", \"MixedCPUsAllocation\", \"MultiArchInstallAzure\", \"NodeSwap\", \"NutanixMultiSubnets\", \"OVNObservability\", \"OpenShiftPodSecurityAdmission\", \"PinnedImages\", \"PlatformOperators\", \"ProcMountType\", \"RouteAdvertisements\", \"RouteExternalCertificate\", \"ServiceAccountTokenNodeBinding\", \"SignatureStores\", \"SigstoreImageVerification\", \"StreamingCollectionEncodingToJSON\", \"StreamingCollectionEncodingToProtobuf\", \"TranslateStreamCloseWebsocketRequests\", \"UpgradeStatus\", \"UserNamespacesPodSecurityStandards\", \"UserNamespacesSupport\", \"VSphereMultiNetworks\", \"VolumeAttributesClass\", \"VolumeGroupSnapshot\"}}","type":"Normal"},{"namespace":"openshift-controller-manager-operator","lastTimestamp":"2025-12-04T11:37:44Z","reason":"ObserveFeatureFlagsUpdated","message":"Updated featureGates to BuildCSIVolumes=true","type":"Normal"},{"namespace":"openshift-controller-manager-operator","lastTimestamp":"2025-12-04T11:37:45Z","reason":"RoleCreated","message":"Created Role.rbac.authorization.k8s.io/system:openshift:sa-creating-openshift-controller-manager -n openshift-infra because it was missing","type":"Normal"},{"namespace":"openshift-controller-manager-operator","lastTimestamp":"2025-12-04T11:37:45Z","reason":"ClusterRoleCreated","message":"Created ClusterRole.rbac.authorization.k8s.io/system:openshift:openshift-controller-manager:image-trigger-controller because it was missing","type":"Normal"},{"namespace":"openshift-controller-manager-operator","lastTimestamp":"2025-12-04T11:37:45Z","reason":"ClusterRoleBindingCreated","message":"Created ClusterRoleBinding.rbac.authorization.k8s.io/system:openshift:openshift-controller-manager:image-trigger-controller because it was missing","type":"Normal"},{"namespace":"openshift-controller-manager-operator","lastTimestamp":"2025-12-04T11:37:45Z","reason":"ClusterRoleBindingCreated","message":"Created ClusterRoleBinding.rbac.authorization.k8s.io/system:deployer because it was missing","type":"Normal"},{"namespace":"openshift-controller-manager-operator","lastTimestamp":"2025-12-04T11:37:45Z","reason":"ClusterRoleCreated","message":"Created ClusterRole.rbac.authorization.k8s.io/system:deployer because it was missing","type":"Normal"},{"namespace":"openshift-controller-manager-operator","lastTimestamp":"2025-12-04T11:37:45Z","reason":"ClusterRoleBindingCreated","message":"Created ClusterRoleBinding.rbac.authorization.k8s.io/system:openshift:tokenreview-openshift-controller-manager because it was missing","type":"Normal"},{"namespace":"openshift-controller-manager-operator","lastTimestamp":"2025-12-04T11:37:45Z","reason":"ConfigMapCreateFailed","message":"Failed to create ConfigMap/config -n openshift-controller-manager: namespaces \"openshift-controller-manager\" not found","type":"Warning"},{"namespace":"openshift-controller-manager-operator","lastTimestamp":"2025-12-04T11:37:45Z","reason":"RoleCreateFailed","message":"Failed to create Role.rbac.authorization.k8s.io/system:openshift:leader-locking-openshift-controller-manager -n openshift-controller-manager: namespaces \"openshift-controller-manager\" not found","type":"Warning"},{"namespace":"openshift-controller-manager-operator","lastTimestamp":"2025-12-04T11:37:45Z","reason":"ConfigMapCreateFailed","message":"Failed to create configmap/openshift-service-ca-n openshift-controller-manager: namespaces \"openshift-controller-manager\" not found","type":"Normal"},{"namespace":"openshift-controller-manager-operator","lastTimestamp":"2025-12-04T11:37:45Z","reason":"RoleBindingCreateFailed","message":"Failed to create RoleBinding.rbac.authorization.k8s.io/system:openshift:leader-locking-openshift-controller-manager -n openshift-controller-manager: namespaces \"openshift-controller-manager\" not found","type":"Warning"},{"namespace":"openshift-controller-manager-operator","lastTimestamp":"2025-12-04T11:37:45Z","reason":"NamespaceCreated","message":"Created Namespace/openshift-controller-manager because it was missing","type":"Normal"},{"namespace":"openshift-controller-manager-operator","lastTimestamp":"2025-12-04T11:37:45Z","reason":"ClusterRoleCreated","message":"Created ClusterRole.rbac.authorization.k8s.io/system:openshift:openshift-route-controller-manager because it was missing","type":"Normal"},{"namespace":"openshift-controller-manager-operator","lastTimestamp":"2025-12-04T11:37:45Z","reason":"ClusterRoleBindingCreated","message":"Created ClusterRoleBinding.rbac.authorization.k8s.io/system:openshift:openshift-route-controller-manager because it was missing","type":"Normal"},{"namespace":"openshift-controller-manager-operator","lastTimestamp":"2025-12-04T11:37:45Z","reason":"ConfigMapCreated","message":"Created ConfigMap/openshift-global-ca -n openshift-controller-manager because it was missing","type":"Normal"},{"namespace":"openshift-controller-manager-operator","lastTimestamp":"2025-12-04T11:37:45Z","reason":"RoleCreateFailed","message":"Failed to create Role.rbac.authorization.k8s.io/system:openshift:leader-locking-openshift-route-controller-manager -n openshift-route-controller-manager: namespaces \"openshift-route-controller-manager\" not found","type":"Warning"},{"namespace":"openshift-controller-manager-operator","lastTimestamp":"2025-12-04T11:37:45Z","reason":"ConfigMapCreateFailed","message":"Failed to create ConfigMap/config -n openshift-route-controller-manager: namespaces \"openshift-route-controller-manager\" not found","type":"Warning"},{"namespace":"openshift-controller-manager-operator","lastTimestamp":"2025-12-04T11:37:45Z","reason":"DeploymentCreated","message":"Created Deployment.apps/controller-manager -n openshift-controller-manager because it was missing","type":"Normal"},{"namespace":"openshift-controller-manager-operator","lastTimestamp":"2025-12-04T11:37:45Z","reason":"RoleBindingCreateFailed","message":"Failed to create RoleBinding.rbac.authorization.k8s.io/system:openshift:leader-locking-openshift-route-controller-manager -n openshift-route-controller-manager: namespaces \"openshift-route-controller-manager\" not found","type":"Warning"},{"namespace":"openshift-controller-manager-operator","lastTimestamp":"2025-12-04T11:37:45Z","reason":"NamespaceCreated","message":"Created Namespace/openshift-route-controller-manager because it was missing","type":"Normal"},{"namespace":"openshift-controller-manager-operator","lastTimestamp":"2025-12-04T11:37:45Z","reason":"ServiceAccountCreated","message":"Created ServiceAccount/route-controller-manager-sa -n openshift-route-controller-manager because it was missing","type":"Normal"},{"namespace":"openshift-controller-manager-operator","lastTimestamp":"2025-12-04T11:37:45Z","reason":"RoleCreated","message":"Created Role.rbac.authorization.k8s.io/system:openshift:sa-creating-route-controller-manager -n openshift-infra because it was missing","type":"Normal"},{"namespace":"openshift-controller-manager-operator","lastTimestamp":"2025-12-04T11:37:45Z","reason":"RoleBindingCreated","message":"Created RoleBinding.rbac.authorization.k8s.io/system:openshift:sa-creating-route-controller-manager -n openshift-infra because it was missing","type":"Normal"},{"namespace":"openshift-controller-manager-operator","lastTimestamp":"2025-12-04T11:37:45Z","reason":"RoleCreated","message":"Created Role.rbac.authorization.k8s.io/prometheus-k8s -n openshift-route-controller-manager because it was missing","type":"Normal"},{"namespace":"openshift-controller-manager-operator","lastTimestamp":"2025-12-04T11:37:45Z","reason":"RoleBindingCreated","message":"Created RoleBinding.rbac.authorization.k8s.io/prometheus-k8s -n openshift-route-controller-manager because it was missing","type":"Normal"},{"namespace":"openshift-controller-manager-operator","lastTimestamp":"2025-12-04T11:37:45Z","reason":"ServiceCreated","message":"Created Service/route-controller-manager -n openshift-route-controller-manager because it was missing","type":"Normal"},{"namespace":"openshift-controller-manager-operator","lastTimestamp":"2025-12-04T11:37:45Z","reason":"ClusterRoleCreated","message":"Created ClusterRole.rbac.authorization.k8s.io/system:openshift:tokenreview-openshift-route-controller-manager because it was missing","type":"Normal"},{"namespace":"openshift-controller-manager-operator","lastTimestamp":"2025-12-04T11:37:45Z","reason":"ClusterRoleBindingCreated","message":"Created ClusterRoleBinding.rbac.authorization.k8s.io/system:openshift:tokenreview-openshift-route-controller-manager because it was missing","type":"Normal"},{"namespace":"openshift-controller-manager-operator","lastTimestamp":"2025-12-04T11:37:45Z","reason":"ClusterRoleCreated","message":"Created ClusterRole.rbac.authorization.k8s.io/system:openshift:openshift-controller-manager:ingress-to-route-controller because it was missing","type":"Normal"},{"namespace":"openshift-controller-manager-operator","lastTimestamp":"2025-12-04T11:37:45Z","reason":"ClusterRoleBindingCreated","message":"Created ClusterRoleBinding.rbac.authorization.k8s.io/system:openshift:openshift-controller-manager:ingress-to-route-controller because it was missing","type":"Normal"},{"namespace":"openshift-controller-manager-operator","lastTimestamp":"2025-12-04T11:37:45Z","reason":"RoleCreated","message":"Created Role.rbac.authorization.k8s.io/system:openshift:leader-locking-openshift-controller-manager -n kube-system because it was missing","type":"Normal"},{"namespace":"openshift-controller-manager-operator","lastTimestamp":"2025-12-04T11:37:45Z","reason":"RoleBindingCreated","message":"Created RoleBinding.rbac.authorization.k8s.io/system:openshift:leader-locking-openshift-controller-manager -n kube-system because it was missing","type":"Normal"},{"namespace":"openshift-controller-manager-operator","lastTimestamp":"2025-12-04T11:37:45Z","reason":"ClusterRoleBindingCreated","message":"Created ClusterRoleBinding.rbac.authorization.k8s.io/system:openshift:openshift-controller-manager:update-buildconfig-status because it was missing","type":"Normal"},{"namespace":"openshift-controller-manager-operator","lastTimestamp":"2025-12-04T11:37:45Z","reason":"RoleBindingCreated","message":"Created RoleBinding.rbac.authorization.k8s.io/system:openshift:sa-creating-openshift-controller-manager -n openshift-infra because it was missing","type":"Normal"},{"namespace":"openshift-controller-manager-operator","lastTimestamp":"2025-12-04T11:37:45Z","reason":"ServiceAccountCreated","message":"Created ServiceAccount/openshift-controller-manager-sa -n openshift-controller-manager because it was missing","type":"Normal"},{"namespace":"openshift-controller-manager-operator","lastTimestamp":"2025-12-04T11:37:45Z","reason":"ServiceCreated","message":"Created Service/controller-manager -n openshift-controller-manager because it was missing","type":"Normal"},{"namespace":"openshift-controller-manager-operator","lastTimestamp":"2025-12-04T11:37:45Z","reason":"RoleCreated","message":"Created Role.rbac.authorization.k8s.io/prometheus-k8s -n openshift-controller-manager because it was missing","type":"Normal"},{"namespace":"openshift-controller-manager-operator","lastTimestamp":"2025-12-04T11:37:45Z","reason":"RoleBindingCreated","message":"Created RoleBinding.rbac.authorization.k8s.io/prometheus-k8s -n openshift-controller-manager because it was missing","type":"Normal"},{"namespace":"openshift-controller-manager-operator","lastTimestamp":"2025-12-04T11:37:45Z","reason":"ClusterRoleCreated","message":"Created ClusterRole.rbac.authorization.k8s.io/system:openshift:openshift-controller-manager:update-buildconfig-status because it was missing","type":"Normal"},{"namespace":"openshift-controller-manager-operator","lastTimestamp":"2025-12-04T11:37:46Z","reason":"OperatorStatusChanged","message":"Status for clusteroperator/openshift-controller-manager changed: Degraded changed from Unknown to False (\"All is well\")","type":"Normal"},{"namespace":"openshift-controller-manager-operator","lastTimestamp":"2025-12-04T11:37:46Z","reason":"DeploymentCreated","message":"Created Deployment.apps/route-controller-manager -n openshift-route-controller-manager because it was missing","type":"Normal"},{"namespace":"openshift-controller-manager-operator","lastTimestamp":"2025-12-04T11:37:47Z","reason":"RoleCreated","message":"Created Role.rbac.authorization.k8s.io/system:openshift:leader-locking-openshift-route-controller-manager -n openshift-route-controller-manager because it was missing","type":"Normal"},{"namespace":"openshift-controller-manager-operator","lastTimestamp":"2025-12-04T11:37:47Z","reason":"RoleCreated","message":"Created Role.rbac.authorization.k8s.io/system:openshift:leader-locking-openshift-controller-manager -n openshift-controller-manager because it was missing","type":"Normal"},{"namespace":"openshift-controller-manager-operator","lastTimestamp":"2025-12-04T11:37:47Z","reason":"RoleBindingCreated","message":"Created RoleBinding.rbac.authorization.k8s.io/system:openshift:leader-locking-openshift-controller-manager -n openshift-controller-manager because it was missing","type":"Normal"},{"namespace":"openshift-controller-manager-operator","lastTimestamp":"2025-12-04T11:37:47Z","reason":"ConfigMapCreated","message":"Created ConfigMap/config -n openshift-route-controller-manager because it was missing","type":"Normal"},{"namespace":"openshift-controller-manager-operator","lastTimestamp":"2025-12-04T11:37:47Z","reason":"RoleBindingCreated","message":"Created RoleBinding.rbac.authorization.k8s.io/system:openshift:leader-locking-openshift-route-controller-manager -n openshift-route-controller-manager because it was missing","type":"Normal"},{"namespace":"openshift-controller-manager-operator","lastTimestamp":"2025-12-04T11:37:47Z","reason":"ConfigMapCreated","message":"Created ConfigMap/config -n openshift-controller-manager because it was missing","type":"Normal"},{"namespace":"openshift-controller-manager-operator","lastTimestamp":"2025-12-04T11:37:47Z","reason":"ConfigMapCreated","message":"Created configmap/openshift-service-ca-n openshift-controller-manager because it was missing","type":"Normal"},{"namespace":"openshift-controller-manager-operator","lastTimestamp":"2025-12-04T11:37:48Z","reason":"OperatorStatusChanged","message":"Status for clusteroperator/openshift-controller-manager changed: Degraded message changed from \"All is well\" to \"OpenshiftControllerManagerStaticResourcesDegraded: \\\"assets/openshift-controller-manager/leader-role.yaml\\\" (string): namespaces \\\"openshift-controller-manager\\\" not found\\nOpenshiftControllerManagerStaticResourcesDegraded: \\\"assets/openshift-controller-manager/leader-rolebinding.yaml\\\" (string): namespaces \\\"openshift-controller-manager\\\" not found\\nOpenshiftControllerManagerStaticResourcesDegraded: \\\"assets/openshift-controller-manager/route-controller-manager-leader-role.yaml\\\" (string): namespaces \\\"openshift-route-controller-manager\\\" not found\\nOpenshiftControllerManagerStaticResourcesDegraded: \\\"assets/openshift-controller-manager/route-controller-manager-leader-rolebinding.yaml\\\" (string): namespaces \\\"openshift-route-controller-manager\\\" not found\\nOpenshiftControllerManagerStaticResourcesDegraded: \"","type":"Normal"},{"namespace":"openshift-controller-manager-operator","lastTimestamp":"2025-12-04T11:37:50Z","reason":"OperatorStatusChanged","message":"Status for clusteroperator/openshift-controller-manager changed: Progressing changed from Unknown to True (\"Progressing: deployment/controller-manager: observed generation is 1, desired generation is 2.\\nProgressing: deployment/controller-manager: available replicas is 0, desired available replicas \u003e 1\\nProgressing: deployment/route-controller-manager: observed generation is 1, desired generation is 2.\\nProgressing: deployment/route-controller-manager: available replicas is 0, desired available replicas \u003e 1\\nProgressing: openshiftcontrollermanagers.operator.openshift.io/cluster: observed generation is 0, desired generation is 2.\"),Available changed from Unknown to False (\"Available: no pods available on any node.\"),Upgradeable changed from Unknown to True (\"All is well\")","type":"Normal"},{"namespace":"openshift-controller-manager-operator","lastTimestamp":"2025-12-04T11:37:51Z","reason":"OperatorStatusChanged","message":"Status for clusteroperator/openshift-controller-manager changed: Degraded message changed from \"OpenshiftControllerManagerStaticResourcesDegraded: \\\"assets/openshift-controller-manager/leader-role.yaml\\\" (string): namespaces \\\"openshift-controller-manager\\\" not found\\nOpenshiftControllerManagerStaticResourcesDegraded: \\\"assets/openshift-controller-manager/leader-rolebinding.yaml\\\" (string): namespaces \\\"openshift-controller-manager\\\" not found\\nOpenshiftControllerManagerStaticResourcesDegraded: \\\"assets/openshift-controller-manager/route-controller-manager-leader-role.yaml\\\" (string): namespaces \\\"openshift-route-controller-manager\\\" not found\\nOpenshiftControllerManagerStaticResourcesDegraded: \\\"assets/openshift-controller-manager/route-controller-manager-leader-rolebinding.yaml\\\" (string): namespaces \\\"openshift-route-controller-manager\\\" not found\\nOpenshiftControllerManagerStaticResourcesDegraded: \" to \"All is well\"","type":"Normal"},{"namespace":"openshift-controller-manager-operator","lastTimestamp":"2025-12-04T11:37:53Z","reason":"OperatorStatusChanged","message":"Status for clusteroperator/openshift-controller-manager changed: Progressing message changed from \"Progressing: deployment/controller-manager: observed generation is 1, desired generation is 2.\\nProgressing: deployment/controller-manager: available replicas is 0, desired available replicas \u003e 1\\nProgressing: deployment/route-controller-manager: observed generation is 1, desired generation is 2.\\nProgressing: deployment/route-controller-manager: available replicas is 0, desired available replicas \u003e 1\\nProgressing: openshiftcontrollermanagers.operator.openshift.io/cluster: observed generation is 0, desired generation is 2.\" to \"Progressing: deployment/controller-manager: available replicas is 0, desired available replicas \u003e 1\\nProgressing: deployment/route-controller-manager: available replicas is 0, desired available replicas \u003e 1\"","type":"Normal"},{"namespace":"openshift-controller-manager-operator","lastTimestamp":"2025-12-04T11:38:05Z","reason":"ConfigMapUpdated","message":"Updated ConfigMap/config -n openshift-controller-manager:\ncause by changes in data.openshift-controller-manager.serving-cert.secret","type":"Normal"},{"namespace":"openshift-controller-manager-operator","lastTimestamp":"2025-12-04T11:38:05Z","reason":"ConfigMapCreated","message":"Created ConfigMap/client-ca -n openshift-controller-manager because it was missing","type":"Normal"},{"namespace":"openshift-controller-manager-operator","lastTimestamp":"2025-12-04T11:38:05Z","reason":"ConfigMapCreated","message":"Created ConfigMap/client-ca -n openshift-route-controller-manager because it was missing","type":"Normal"},{"namespace":"openshift-controller-manager-operator","lastTimestamp":"2025-12-04T11:38:10Z","reason":"OperatorStatusChanged","message":"Status for clusteroperator/openshift-controller-manager changed: Progressing message changed from \"Progressing: deployment/controller-manager: available replicas is 0, desired available replicas \u003e 1\\nProgressing: deployment/route-controller-manager: available replicas is 0, desired available replicas \u003e 1\" to \"Progressing: deployment/controller-manager: observed generation is 2, desired generation is 3.\\nProgressing: deployment/controller-manager: available replicas is 0, desired available replicas \u003e 1\\nProgressing: deployment/route-controller-manager: observed generation is 2, desired generation is 3.\\nProgressing: deployment/route-controller-manager: available replicas is 0, desired available replicas \u003e 1\"","type":"Normal"},{"namespace":"openshift-controller-manager-operator","lastTimestamp":"2025-12-04T11:38:24Z","reason":"OperatorStatusChanged","message":"Status for clusteroperator/openshift-controller-manager changed: Progressing message changed from \"Progressing: deployment/controller-manager: observed generation is 2, desired generation is 3.\\nProgressing: deployment/controller-manager: available replicas is 0, desired available replicas \u003e 1\\nProgressing: deployment/route-controller-manager: observed generation is 2, desired generation is 3.\\nProgressing: deployment/route-controller-manager: available replicas is 0, desired available replicas \u003e 1\" to \"Progressing: deployment/controller-manager: observed generation is 3, desired generation is 4.\\nProgressing: deployment/controller-manager: available replicas is 0, desired available replicas \u003e 1\\nProgressing: deployment/route-controller-manager: observed generation is 3, desired generation is 4.\\nProgressing: deployment/route-controller-manager: available replicas is 0, desired available replicas \u003e 1\"","type":"Normal"},{"namespace":"openshift-controller-manager-operator","lastTimestamp":"2025-12-04T11:38:24Z","reason":"ConfigMapUpdated","message":"Updated ConfigMap/client-ca -n openshift-controller-manager:\ncause by changes in data.ca-bundle.crt","type":"Normal"},{"namespace":"openshift-controller-manager-operator","lastTimestamp":"2025-12-04T11:38:24Z","reason":"ConfigMapUpdated","message":"Updated ConfigMap/client-ca -n openshift-route-controller-manager:\ncause by changes in data.ca-bundle.crt","type":"Normal"},{"namespace":"openshift-controller-manager-operator","lastTimestamp":"2025-12-04T11:38:24Z","reason":"ConfigMapUpdated","message":"Updated ConfigMap/config -n openshift-route-controller-manager:\ncause by changes in data.openshift-route-controller-manager.client-ca.configmap,data.openshift-route-controller-manager.serving-cert.secret","type":"Normal"},{"namespace":"openshift-controller-manager-operator","lastTimestamp":"2025-12-04T11:38:45Z","reason":"ConfigMapUpdated","message":"Updated ConfigMap/config -n openshift-route-controller-manager:\ncause by changes in data.openshift-route-controller-manager.client-ca.configmap","type":"Normal"},{"namespace":"openshift-controller-manager-operator","lastTimestamp":"2025-12-04T11:38:45Z","reason":"DeploymentUpdated","message":"Updated Deployment.apps/controller-manager -n openshift-controller-manager because it changed","type":"Normal"},{"namespace":"openshift-controller-manager-operator","lastTimestamp":"2025-12-04T11:38:45Z","reason":"DeploymentUpdated","message":"Updated Deployment.apps/route-controller-manager -n openshift-route-controller-manager because it changed","type":"Normal"},{"namespace":"openshift-controller-manager-operator","lastTimestamp":"2025-12-04T11:38:45Z","reason":"ConfigMapUpdated","message":"Updated ConfigMap/config -n openshift-controller-manager:\ncause by changes in data.openshift-controller-manager.client-ca.configmap","type":"Normal"},{"namespace":"openshift-controller-manager-operator","lastTimestamp":"2025-12-04T11:38:46Z","reason":"OperatorVersionChanged","message":"clusteroperator/openshift-controller-manager version \"operator\" changed from \"\" to \"4.18.29\"","type":"Normal"},{"namespace":"openshift-controller-manager-operator","lastTimestamp":"2025-12-04T11:38:47Z","reason":"OperatorStatusChanged","message":"Status for clusteroperator/openshift-controller-manager changed: Progressing message changed from \"Progressing: deployment/controller-manager: observed generation is 3, desired generation is 4.\\nProgressing: deployment/controller-manager: available replicas is 0, desired available replicas \u003e 1\\nProgressing: deployment/route-controller-manager: observed generation is 3, desired generation is 4.\\nProgressing: deployment/route-controller-manager: available replicas is 0, desired available replicas \u003e 1\" to \"Progressing: deployment/controller-manager: observed generation is 4, desired generation is 5.\\nProgressing: deployment/route-controller-manager: observed generation is 4, desired generation is 5.\",Available changed from False to True (\"All is well\"),status.versions changed from [] to [{\"operator\" \"4.18.29\"}]","type":"Normal"},{"namespace":"openshift-controller-manager-operator","lastTimestamp":"2025-12-04T11:43:19Z","reason":"BackOff","message":"Back-off restarting failed container openshift-controller-manager-operator in pod openshift-controller-manager-operator-6c8676f99d-7z948_openshift-controller-manager-operator(3322cc5a-f1f7-4522-b423-19fb7f38cd43)","type":"Warning"},{"namespace":"openshift-controller-manager-operator","lastTimestamp":"2025-12-04T11:50:40Z","reason":"LeaderElection","message":"openshift-controller-manager-operator-6c8676f99d-7z948_8b885916-2eb8-4873-93e8-1f63051ec480 became leader","type":"Normal"},{"namespace":"openshift-controller-manager-operator","lastTimestamp":"2025-12-04T11:50:40Z","reason":"FeatureGatesInitialized","message":"FeatureGates updated to featuregates.Features{Enabled:[]v1.FeatureGateName{\"AWSEFSDriverVolumeMetrics\", \"AdminNetworkPolicy\", \"AlibabaPlatform\", \"AzureWorkloadIdentity\", \"BareMetalLoadBalancer\", \"BuildCSIVolumes\", \"ChunkSizeMiB\", \"CloudDualStackNodeIPs\", \"DisableKubeletCloudCredentialProviders\", \"GCPLabelsTags\", \"HardwareSpeed\", \"IngressControllerLBSubnetsAWS\", \"KMSv1\", \"ManagedBootImages\", \"ManagedBootImagesAWS\", \"MultiArchInstallAWS\", \"MultiArchInstallGCP\", \"NetworkDiagnosticsConfig\", \"NetworkLiveMigration\", \"NetworkSegmentation\", \"NewOLM\", \"NodeDisruptionPolicy\", \"OnClusterBuild\", \"PersistentIPsForVirtualization\", \"PrivateHostedZoneAWS\", \"SetEIPForNLBIngressController\", \"VSphereControlPlaneMachineSet\", \"VSphereDriverConfiguration\", \"VSphereMultiVCenters\", \"VSphereStaticIPs\", \"ValidatingAdmissionPolicy\"}, Disabled:[]v1.FeatureGateName{\"AWSClusterHostedDNS\", \"AdditionalRoutingCapabilities\", \"AutomatedEtcdBackup\", \"BootcNodeManagement\", \"CSIDriverSharedResource\", \"ClusterAPIInstall\", \"ClusterAPIInstallIBMCloud\", \"ClusterMonitoringConfig\", \"ConsolePluginContentSecurityPolicy\", \"DNSNameResolver\", \"DynamicResourceAllocation\", \"EtcdBackendQuota\", \"EventedPLEG\", \"Example\", \"ExternalOIDC\", \"ExternalOIDCWithUIDAndExtraClaimMappings\", \"GCPClusterHostedDNS\", \"GatewayAPI\", \"ImageStreamImportMode\", \"IngressControllerDynamicConfigurationManager\", \"InsightsConfig\", \"InsightsConfigAPI\", \"InsightsOnDemandDataGather\", \"InsightsRuntimeExtractor\", \"MachineAPIMigration\", \"MachineAPIOperatorDisableMachineHealthCheckController\", \"MachineAPIProviderOpenStack\", \"MachineConfigNodes\", \"MaxUnavailableStatefulSet\", \"MetricsCollectionProfiles\", \"MinimumKubeletVersion\", \"MixedCPUsAllocation\", \"MultiArchInstallAzure\", \"NodeSwap\", \"NutanixMultiSubnets\", \"OVNObservability\", \"OpenShiftPodSecurityAdmission\", \"PinnedImages\", \"PlatformOperators\", \"ProcMountType\", \"RouteAdvertisements\", \"RouteExternalCertificate\", \"ServiceAccountTokenNodeBinding\", \"SignatureStores\", \"SigstoreImageVerification\", \"StreamingCollectionEncodingToJSON\", \"StreamingCollectionEncodingToProtobuf\", \"TranslateStreamCloseWebsocketRequests\", \"UpgradeStatus\", \"UserNamespacesPodSecurityStandards\", \"UserNamespacesSupport\", \"VSphereMultiNetworks\", \"VolumeAttributesClass\", \"VolumeGroupSnapshot\"}}","type":"Normal"},{"namespace":"openshift-controller-manager-operator","lastTimestamp":"2025-12-04T11:50:40Z","reason":"OperatorStatusChanged","message":"Status for clusteroperator/openshift-controller-manager changed: Progressing message changed from \"Progressing: deployment/controller-manager: observed generation is 4, desired generation is 5.\\nProgressing: deployment/route-controller-manager: observed generation is 4, desired generation is 5.\" to \"Progressing: deployment/controller-manager: available replicas is 0, desired available replicas \u003e 1\\nProgressing: deployment/route-controller-manager: available replicas is 0, desired available replicas \u003e 1\",Available changed from True to False (\"Available: no pods available on any node.\")","type":"Normal"},{"namespace":"openshift-controller-manager-operator","lastTimestamp":"2025-12-04T11:55:47Z","reason":"Created","message":"Created container: openshift-controller-manager-operator","type":"Normal"},{"namespace":"openshift-controller-manager-operator","lastTimestamp":"2025-12-04T11:55:47Z","reason":"Started","message":"Started container openshift-controller-manager-operator","type":"Normal"},{"namespace":"openshift-controller-manager-operator","lastTimestamp":"2025-12-04T11:55:47Z","reason":"Pulled","message":"Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8eabac819f289e29d75c7ab172d8124554849a47f0b00770928c3eb19a5a31c4\" already present on machine","type":"Normal"},{"namespace":"openshift-controller-manager-operator","lastTimestamp":"2025-12-04T11:56:26Z","reason":"LeaderElection","message":"openshift-controller-manager-operator-6c8676f99d-7z948_8ba4cf9c-e623-440c-8949-fbff56d5d895 became leader","type":"Normal"},{"namespace":"openshift-controller-manager-operator","lastTimestamp":"2025-12-04T11:56:26Z","reason":"FeatureGatesInitialized","message":"FeatureGates updated to featuregates.Features{Enabled:[]v1.FeatureGateName{\"AWSEFSDriverVolumeMetrics\", \"AdminNetworkPolicy\", \"AlibabaPlatform\", \"AzureWorkloadIdentity\", \"BareMetalLoadBalancer\", \"BuildCSIVolumes\", \"ChunkSizeMiB\", \"CloudDualStackNodeIPs\", \"DisableKubeletCloudCredentialProviders\", \"GCPLabelsTags\", \"HardwareSpeed\", \"IngressControllerLBSubnetsAWS\", \"KMSv1\", \"ManagedBootImages\", \"ManagedBootImagesAWS\", \"MultiArchInstallAWS\", \"MultiArchInstallGCP\", \"NetworkDiagnosticsConfig\", \"NetworkLiveMigration\", \"NetworkSegmentation\", \"NewOLM\", \"NodeDisruptionPolicy\", \"OnClusterBuild\", \"PersistentIPsForVirtualization\", \"PrivateHostedZoneAWS\", \"SetEIPForNLBIngressController\", \"VSphereControlPlaneMachineSet\", \"VSphereDriverConfiguration\", \"VSphereMultiVCenters\", \"VSphereStaticIPs\", \"ValidatingAdmissionPolicy\"}, Disabled:[]v1.FeatureGateName{\"AWSClusterHostedDNS\", \"AdditionalRoutingCapabilities\", \"AutomatedEtcdBackup\", \"BootcNodeManagement\", \"CSIDriverSharedResource\", \"ClusterAPIInstall\", \"ClusterAPIInstallIBMCloud\", \"ClusterMonitoringConfig\", \"ConsolePluginContentSecurityPolicy\", \"DNSNameResolver\", \"DynamicResourceAllocation\", \"EtcdBackendQuota\", \"EventedPLEG\", \"Example\", \"ExternalOIDC\", \"ExternalOIDCWithUIDAndExtraClaimMappings\", \"GCPClusterHostedDNS\", \"GatewayAPI\", \"ImageStreamImportMode\", \"IngressControllerDynamicConfigurationManager\", \"InsightsConfig\", \"InsightsConfigAPI\", \"InsightsOnDemandDataGather\", \"InsightsRuntimeExtractor\", \"MachineAPIMigration\", \"MachineAPIOperatorDisableMachineHealthCheckController\", \"MachineAPIProviderOpenStack\", \"MachineConfigNodes\", \"MaxUnavailableStatefulSet\", \"MetricsCollectionProfiles\", \"MinimumKubeletVersion\", \"MixedCPUsAllocation\", \"MultiArchInstallAzure\", \"NodeSwap\", \"NutanixMultiSubnets\", \"OVNObservability\", \"OpenShiftPodSecurityAdmission\", \"PinnedImages\", \"PlatformOperators\", \"ProcMountType\", \"RouteAdvertisements\", \"RouteExternalCertificate\", \"ServiceAccountTokenNodeBinding\", \"SignatureStores\", \"SigstoreImageVerification\", \"StreamingCollectionEncodingToJSON\", \"StreamingCollectionEncodingToProtobuf\", \"TranslateStreamCloseWebsocketRequests\", \"UpgradeStatus\", \"UserNamespacesPodSecurityStandards\", \"UserNamespacesSupport\", \"VSphereMultiNetworks\", \"VolumeAttributesClass\", \"VolumeGroupSnapshot\"}}","type":"Normal"},{"namespace":"openshift-controller-manager-operator","lastTimestamp":"2025-12-04T12:00:26Z","reason":"ObservedConfigChanged","message":"Writing updated observed config:   map[string]any{\n  \t\"build\": map[string]any{\"buildDefaults\": map[string]any{\"resources\": map[string]any{}}, \"imageTemplateFormat\": map[string]any{\"format\": string(\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:31aa3c7464\"...)}},\n  \t\"controllers\": []any{\n  \t\t... // 8 identical elements\n  \t\tstring(\"openshift.io/deploymentconfig\"),\n  \t\tstring(\"openshift.io/image-import\"),\n  \t\tstrings.Join({\n+ \t\t\t\"-\",\n  \t\t\t\"openshift.io/image-puller-rolebindings\",\n  \t\t}, \"\"),\n  \t\tstring(\"openshift.io/image-signature-import\"),\n  \t\tstring(\"openshift.io/image-trigger\"),\n  \t\t... // 2 identical elements\n  \t\tstring(\"openshift.io/origin-namespace\"),\n  \t\tstring(\"openshift.io/serviceaccount\"),\n  \t\tstrings.Join({\n+ \t\t\t\"-\",\n  \t\t\t\"openshift.io/serviceaccount-pull-secrets\",\n  \t\t}, \"\"),\n  \t\tstring(\"openshift.io/templateinstance\"),\n  \t\tstring(\"openshift.io/templateinstancefinalizer\"),\n  \t\tstring(\"openshift.io/unidling\"),\n  \t},\n  \t\"deployer\": map[string]any{\"imageTemplateFormat\": map[string]any{\"format\": string(\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:42c3f5030d\"...)}},\n  \t\"featureGates\": []any{string(\"BuildCSIVolumes=true\")},\n  \t\"ingress\": map[string]any{\"ingressIPNetworkCIDR\": string(\"\")},\n  }\n","type":"Normal"},{"namespace":"openshift-controller-manager-operator","lastTimestamp":"2025-12-04T12:00:29Z","reason":"ConfigMapUpdated","message":"Updated ConfigMap/config -n openshift-controller-manager:\ncause by changes in data.config.yaml","type":"Normal"},{"namespace":"openshift-controller-manager-operator","lastTimestamp":"2025-12-04T12:00:29Z","reason":"ConfigMapUpdated","message":"Updated ConfigMap/config -n openshift-route-controller-manager:\ncause by changes in data.config.yaml","type":"Normal"},{"namespace":"openshift-controller-manager-operator","lastTimestamp":"2025-12-04T12:00:29Z","reason":"DeploymentUpdated","message":"Updated Deployment.apps/controller-manager -n openshift-controller-manager because it changed","type":"Normal"},{"namespace":"openshift-controller-manager-operator","lastTimestamp":"2025-12-04T12:00:29Z","reason":"DeploymentUpdated","message":"Updated Deployment.apps/route-controller-manager -n openshift-route-controller-manager because it changed","type":"Normal"},{"namespace":"openshift-controller-manager-operator","lastTimestamp":"2025-12-04T12:00:29Z","reason":"OperatorStatusChanged","message":"Status for clusteroperator/openshift-controller-manager changed: Progressing message changed from \"Progressing: deployment/controller-manager: available replicas is 0, desired available replicas \u003e 1\\nProgressing: deployment/route-controller-manager: available replicas is 0, desired available replicas \u003e 1\" to \"Progressing: deployment/controller-manager: observed generation is 5, desired generation is 6.\\nProgressing: deployment/controller-manager: available replicas is 0, desired available replicas \u003e 1\\nProgressing: deployment/route-controller-manager: observed generation is 5, desired generation is 6.\\nProgressing: deployment/route-controller-manager: available replicas is 0, desired available replicas \u003e 1\\nProgressing: openshiftcontrollermanagers.operator.openshift.io/cluster: observed generation is 2, desired generation is 3.\"","type":"Normal"},{"namespace":"openshift-controller-manager-operator","lastTimestamp":"2025-12-04T12:00:49Z","reason":"OperatorStatusChanged","message":"Status for clusteroperator/openshift-controller-manager changed: Progressing message changed from \"Progressing: deployment/controller-manager: observed generation is 5, desired generation is 6.\\nProgressing: deployment/controller-manager: available replicas is 0, desired available replicas \u003e 1\\nProgressing: deployment/route-controller-manager: observed generation is 5, desired generation is 6.\\nProgressing: deployment/route-controller-manager: available replicas is 0, desired available replicas \u003e 1\\nProgressing: openshiftcontrollermanagers.operator.openshift.io/cluster: observed generation is 2, desired generation is 3.\" to \"Progressing: deployment/controller-manager: observed generation is 5, desired generation is 6.\\nProgressing: deployment/controller-manager: available replicas is 0, desired available replicas \u003e 1\\nProgressing: deployment/route-controller-manager: observed generation is 5, desired generation is 6.\\nProgressing: deployment/route-controller-manager: available replicas is 0, desired available replicas \u003e 1\"","type":"Normal"}]}events/openshift-monitoring.json0000640000000000000000000002311415114274105015622 0ustar0000000000000000{"items":[{"namespace":"openshift-monitoring","lastTimestamp":"2025-12-04T11:31:42Z","reason":"ScalingReplicaSet","message":"Scaled up replica set cluster-monitoring-operator-7ff994598c to 1","type":"Normal"},{"namespace":"openshift-monitoring","lastTimestamp":"2025-12-04T11:31:45Z","reason":"FailedCreate","message":"Error creating: pods \"cluster-monitoring-operator-7ff994598c-\" is forbidden: autoscaling.openshift.io/ManagementCPUsOverride the cluster does not have any nodes","type":"Warning"},{"namespace":"openshift-monitoring","lastTimestamp":"2025-12-04T11:35:47Z","reason":"FailedCreate","message":"Error creating: pods \"cluster-monitoring-operator-7ff994598c-\" is forbidden: autoscaling.openshift.io/ManagementCPUsOverride the cluster does not have any nodes","type":"Warning"},{"namespace":"openshift-monitoring","lastTimestamp":"2025-12-04T11:35:49Z","reason":"FailedScheduling","message":"0/1 nodes are available: 1 node(s) had untolerated taint {node.kubernetes.io/not-ready: }. preemption: 0/1 nodes are available: 1 Preemption is not helpful for scheduling.","type":"Warning"},{"namespace":"openshift-monitoring","lastTimestamp":"2025-12-04T11:35:49Z","reason":"SuccessfulCreate","message":"Created pod: cluster-monitoring-operator-7ff994598c-p82nn","type":"Normal"},{"namespace":"openshift-monitoring","lastTimestamp":"2025-12-04T11:37:30Z","reason":"Scheduled","message":"Successfully assigned openshift-monitoring/cluster-monitoring-operator-7ff994598c-p82nn to master-0","type":"Normal"},{"namespace":"openshift-monitoring","lastTimestamp":"2025-12-04T11:37:38Z","reason":"FailedMount","message":"MountVolume.SetUp failed for volume \"cluster-monitoring-operator-tls\" : secret \"cluster-monitoring-operator-tls\" not found","type":"Warning"},{"namespace":"openshift-monitoring","lastTimestamp":"2025-12-04T11:38:02Z","reason":"FailedMount","message":"MountVolume.SetUp failed for volume \"cluster-monitoring-operator-tls\" : secret \"cluster-monitoring-operator-tls\" not found","type":"Warning"},{"namespace":"openshift-monitoring","lastTimestamp":"2025-12-04T11:38:19Z","reason":"Pulling","message":"Pulling image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3a77aa4d03b89ea284e3467a268e5989a77a2ef63e685eb1d5c5ea5b3922b7a\"","type":"Normal"},{"namespace":"openshift-monitoring","lastTimestamp":"2025-12-04T11:38:19Z","reason":"AddedInterface","message":"Add eth0 [10.128.0.12/23] from ovn-kubernetes","type":"Normal"},{"namespace":"openshift-monitoring","lastTimestamp":"2025-12-04T11:38:32Z","reason":"Started","message":"Started container cluster-monitoring-operator","type":"Normal"},{"namespace":"openshift-monitoring","lastTimestamp":"2025-12-04T11:38:32Z","reason":"Created","message":"Created container: cluster-monitoring-operator","type":"Normal"},{"namespace":"openshift-monitoring","lastTimestamp":"2025-12-04T11:38:32Z","reason":"Pulled","message":"Successfully pulled image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3a77aa4d03b89ea284e3467a268e5989a77a2ef63e685eb1d5c5ea5b3922b7a\" in 12.556s (12.556s including waiting). Image size: 478917802 bytes.","type":"Normal"},{"namespace":"openshift-monitoring","lastTimestamp":"2025-12-04T11:38:32Z","reason":"FeatureGatesInitialized","message":"FeatureGates updated to featuregates.Features{Enabled:[]v1.FeatureGateName{\"AWSEFSDriverVolumeMetrics\", \"AdminNetworkPolicy\", \"AlibabaPlatform\", \"AzureWorkloadIdentity\", \"BareMetalLoadBalancer\", \"BuildCSIVolumes\", \"ChunkSizeMiB\", \"CloudDualStackNodeIPs\", \"DisableKubeletCloudCredentialProviders\", \"GCPLabelsTags\", \"HardwareSpeed\", \"IngressControllerLBSubnetsAWS\", \"KMSv1\", \"ManagedBootImages\", \"ManagedBootImagesAWS\", \"MultiArchInstallAWS\", \"MultiArchInstallGCP\", \"NetworkDiagnosticsConfig\", \"NetworkLiveMigration\", \"NetworkSegmentation\", \"NewOLM\", \"NodeDisruptionPolicy\", \"OnClusterBuild\", \"PersistentIPsForVirtualization\", \"PrivateHostedZoneAWS\", \"SetEIPForNLBIngressController\", \"VSphereControlPlaneMachineSet\", \"VSphereDriverConfiguration\", \"VSphereMultiVCenters\", \"VSphereStaticIPs\", \"ValidatingAdmissionPolicy\"}, Disabled:[]v1.FeatureGateName{\"AWSClusterHostedDNS\", \"AdditionalRoutingCapabilities\", \"AutomatedEtcdBackup\", \"BootcNodeManagement\", \"CSIDriverSharedResource\", \"ClusterAPIInstall\", \"ClusterAPIInstallIBMCloud\", \"ClusterMonitoringConfig\", \"ConsolePluginContentSecurityPolicy\", \"DNSNameResolver\", \"DynamicResourceAllocation\", \"EtcdBackendQuota\", \"EventedPLEG\", \"Example\", \"ExternalOIDC\", \"ExternalOIDCWithUIDAndExtraClaimMappings\", \"GCPClusterHostedDNS\", \"GatewayAPI\", \"ImageStreamImportMode\", \"IngressControllerDynamicConfigurationManager\", \"InsightsConfig\", \"InsightsConfigAPI\", \"InsightsOnDemandDataGather\", \"InsightsRuntimeExtractor\", \"MachineAPIMigration\", \"MachineAPIOperatorDisableMachineHealthCheckController\", \"MachineAPIProviderOpenStack\", \"MachineConfigNodes\", \"MaxUnavailableStatefulSet\", \"MetricsCollectionProfiles\", \"MinimumKubeletVersion\", \"MixedCPUsAllocation\", \"MultiArchInstallAzure\", \"NodeSwap\", \"NutanixMultiSubnets\", \"OVNObservability\", \"OpenShiftPodSecurityAdmission\", \"PinnedImages\", \"PlatformOperators\", \"ProcMountType\", \"RouteAdvertisements\", \"RouteExternalCertificate\", \"ServiceAccountTokenNodeBinding\", \"SignatureStores\", \"SigstoreImageVerification\", \"StreamingCollectionEncodingToJSON\", \"StreamingCollectionEncodingToProtobuf\", \"TranslateStreamCloseWebsocketRequests\", \"UpgradeStatus\", \"UserNamespacesPodSecurityStandards\", \"UserNamespacesSupport\", \"VSphereMultiNetworks\", \"VolumeAttributesClass\", \"VolumeGroupSnapshot\"}}","type":"Normal"},{"namespace":"openshift-monitoring","lastTimestamp":"2025-12-04T11:38:33Z","reason":"NoValidCertificateFound","message":"No valid client certificate for OpenShiftMonitoringClientCertRequester is found: unable to parse certificate: data does not contain any valid RSA or ECDSA certificates","type":"Normal"},{"namespace":"openshift-monitoring","lastTimestamp":"2025-12-04T11:38:33Z","reason":"ClusterRoleCreated","message":"Created ClusterRole.rbac.authorization.k8s.io/prometheus-operator because it was missing","type":"Normal"},{"namespace":"openshift-monitoring","lastTimestamp":"2025-12-04T11:38:33Z","reason":"CSRCreated","message":"A csr \"system:openshift:openshift-monitoring-62r5r\" is created for OpenShiftMonitoringClientCertRequester","type":"Normal"},{"namespace":"openshift-monitoring","lastTimestamp":"2025-12-04T11:38:33Z","reason":"CSRCreated","message":"A csr \"system:openshift:openshift-monitoring-6959d\" is created for OpenShiftMonitoringTelemeterClientCertRequester","type":"Normal"},{"namespace":"openshift-monitoring","lastTimestamp":"2025-12-04T11:38:33Z","reason":"SecretCreated","message":"Created Secret/alert-relabel-configs -n openshift-monitoring because it was missing","type":"Normal"},{"namespace":"openshift-monitoring","lastTimestamp":"2025-12-04T11:38:33Z","reason":"ServiceAccountCreated","message":"Created ServiceAccount/prometheus-operator -n openshift-monitoring because it was missing","type":"Normal"},{"namespace":"openshift-monitoring","lastTimestamp":"2025-12-04T11:38:33Z","reason":"ClientCertificateCreated","message":"A new client certificate for OpenShiftMonitoringClientCertRequester is available","type":"Normal"},{"namespace":"openshift-monitoring","lastTimestamp":"2025-12-04T11:38:33Z","reason":"ConfigMapCreated","message":"Created ConfigMap/metrics-client-ca -n openshift-monitoring because it was missing","type":"Normal"},{"namespace":"openshift-monitoring","lastTimestamp":"2025-12-04T11:38:33Z","reason":"ClientCertificateCreated","message":"A new client certificate for OpenShiftMonitoringTelemeterClientCertRequester is available","type":"Normal"},{"namespace":"openshift-monitoring","lastTimestamp":"2025-12-04T11:38:33Z","reason":"NoValidCertificateFound","message":"No valid client certificate for OpenShiftMonitoringTelemeterClientCertRequester is found: unable to parse certificate: data does not contain any valid RSA or ECDSA certificates","type":"Normal"},{"namespace":"openshift-monitoring","lastTimestamp":"2025-12-04T11:38:33Z","reason":"ClusterRoleBindingCreated","message":"Created ClusterRoleBinding.rbac.authorization.k8s.io/prometheus-operator because it was missing","type":"Normal"},{"namespace":"openshift-monitoring","lastTimestamp":"2025-12-04T11:38:33Z","reason":"ServiceAccountCreated","message":"Created ServiceAccount/prometheus-operator-admission-webhook -n openshift-monitoring because it was missing","type":"Normal"},{"namespace":"openshift-monitoring","lastTimestamp":"2025-12-04T11:38:33Z","reason":"ServiceCreated","message":"Created Service/prometheus-operator-admission-webhook -n openshift-monitoring because it was missing","type":"Normal"},{"namespace":"openshift-monitoring","lastTimestamp":"2025-12-04T11:38:33Z","reason":"ScalingReplicaSet","message":"Scaled up replica set prometheus-operator-admission-webhook-7c85c4dffd to 1","type":"Normal"},{"namespace":"openshift-monitoring","lastTimestamp":"2025-12-04T11:38:33Z","reason":"FailedScheduling","message":"0/1 nodes are available: 1 node(s) had untolerated taint {node-role.kubernetes.io/master: }. preemption: 0/1 nodes are available: 1 Preemption is not helpful for scheduling.","type":"Warning"},{"namespace":"openshift-monitoring","lastTimestamp":"2025-12-04T11:38:33Z","reason":"SuccessfulCreate","message":"Created pod: prometheus-operator-admission-webhook-7c85c4dffd-xv2wn","type":"Normal"},{"namespace":"openshift-monitoring","lastTimestamp":"2025-12-04T12:01:16Z","reason":"ConfigMapUpdated","message":"Updated ConfigMap/metrics-client-ca -n openshift-monitoring:\ncause by changes in data.client-ca.crt","type":"Normal"}]}events/openshift-machine-config-operator.json0000640000000000000000000003375115114274105020145 0ustar0000000000000000{"items":[{"namespace":"openshift-machine-config-operator","lastTimestamp":"2025-12-04T11:35:53Z","reason":"BackOff","message":"Back-off restarting failed container kube-rbac-proxy-crio in pod kube-rbac-proxy-crio-master-0_openshift-machine-config-operator(3169f44496ed8a28c6d6a15511ab0eec)","type":"Warning"},{"namespace":"openshift-machine-config-operator","lastTimestamp":"2025-12-04T11:36:06Z","reason":"Created","message":"Created container: kube-rbac-proxy-crio","type":"Normal"},{"namespace":"openshift-machine-config-operator","lastTimestamp":"2025-12-04T11:36:06Z","reason":"Started","message":"Started container kube-rbac-proxy-crio","type":"Normal"},{"namespace":"openshift-machine-config-operator","lastTimestamp":"2025-12-04T11:36:06Z","reason":"Pulled","message":"Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69ffd8f8dcceedc2d6eb306cea33f8beabc1be1308cd5f4ee8b9a8e3eab9843\" already present on machine","type":"Normal"},{"namespace":"openshift-machine-config-operator","lastTimestamp":"2025-12-04T11:38:43Z","reason":"SuccessfulCreate","message":"Created pod: machine-config-operator-dc5d7666f-p2cmn","type":"Normal"},{"namespace":"openshift-machine-config-operator","lastTimestamp":"2025-12-04T11:38:43Z","reason":"ScalingReplicaSet","message":"Scaled up replica set machine-config-operator-dc5d7666f to 1","type":"Normal"},{"namespace":"openshift-machine-config-operator","lastTimestamp":"2025-12-04T11:38:44Z","reason":"Scheduled","message":"Successfully assigned openshift-machine-config-operator/machine-config-operator-dc5d7666f-p2cmn to master-0","type":"Normal"},{"namespace":"openshift-machine-config-operator","lastTimestamp":"2025-12-04T11:38:44Z","reason":"AddedInterface","message":"Add eth0 [10.128.0.61/23] from ovn-kubernetes","type":"Normal"},{"namespace":"openshift-machine-config-operator","lastTimestamp":"2025-12-04T11:39:21Z","reason":"Pulled","message":"Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69ffd8f8dcceedc2d6eb306cea33f8beabc1be1308cd5f4ee8b9a8e3eab9843\" already present on machine","type":"Normal"},{"namespace":"openshift-machine-config-operator","lastTimestamp":"2025-12-04T11:39:21Z","reason":"Created","message":"Created container: kube-rbac-proxy","type":"Normal"},{"namespace":"openshift-machine-config-operator","lastTimestamp":"2025-12-04T11:39:22Z","reason":"Started","message":"Started container kube-rbac-proxy","type":"Normal"},{"namespace":"openshift-machine-config-operator","lastTimestamp":"2025-12-04T11:45:10Z","reason":"FeatureGatesInitialized","message":"FeatureGates updated to featuregates.Features{Enabled:[]v1.FeatureGateName{\"AWSEFSDriverVolumeMetrics\", \"AdminNetworkPolicy\", \"AlibabaPlatform\", \"AzureWorkloadIdentity\", \"BareMetalLoadBalancer\", \"BuildCSIVolumes\", \"ChunkSizeMiB\", \"CloudDualStackNodeIPs\", \"DisableKubeletCloudCredentialProviders\", \"GCPLabelsTags\", \"HardwareSpeed\", \"IngressControllerLBSubnetsAWS\", \"KMSv1\", \"ManagedBootImages\", \"ManagedBootImagesAWS\", \"MultiArchInstallAWS\", \"MultiArchInstallGCP\", \"NetworkDiagnosticsConfig\", \"NetworkLiveMigration\", \"NetworkSegmentation\", \"NewOLM\", \"NodeDisruptionPolicy\", \"OnClusterBuild\", \"PersistentIPsForVirtualization\", \"PrivateHostedZoneAWS\", \"SetEIPForNLBIngressController\", \"VSphereControlPlaneMachineSet\", \"VSphereDriverConfiguration\", \"VSphereMultiVCenters\", \"VSphereStaticIPs\", \"ValidatingAdmissionPolicy\"}, Disabled:[]v1.FeatureGateName{\"AWSClusterHostedDNS\", \"AdditionalRoutingCapabilities\", \"AutomatedEtcdBackup\", \"BootcNodeManagement\", \"CSIDriverSharedResource\", \"ClusterAPIInstall\", \"ClusterAPIInstallIBMCloud\", \"ClusterMonitoringConfig\", \"ConsolePluginContentSecurityPolicy\", \"DNSNameResolver\", \"DynamicResourceAllocation\", \"EtcdBackendQuota\", \"EventedPLEG\", \"Example\", \"ExternalOIDC\", \"ExternalOIDCWithUIDAndExtraClaimMappings\", \"GCPClusterHostedDNS\", \"GatewayAPI\", \"ImageStreamImportMode\", \"IngressControllerDynamicConfigurationManager\", \"InsightsConfig\", \"InsightsConfigAPI\", \"InsightsOnDemandDataGather\", \"InsightsRuntimeExtractor\", \"MachineAPIMigration\", \"MachineAPIOperatorDisableMachineHealthCheckController\", \"MachineAPIProviderOpenStack\", \"MachineConfigNodes\", \"MaxUnavailableStatefulSet\", \"MetricsCollectionProfiles\", \"MinimumKubeletVersion\", \"MixedCPUsAllocation\", \"MultiArchInstallAzure\", \"NodeSwap\", \"NutanixMultiSubnets\", \"OVNObservability\", \"OpenShiftPodSecurityAdmission\", \"PinnedImages\", \"PlatformOperators\", \"ProcMountType\", \"RouteAdvertisements\", \"RouteExternalCertificate\", \"ServiceAccountTokenNodeBinding\", \"SignatureStores\", \"SigstoreImageVerification\", \"StreamingCollectionEncodingToJSON\", \"StreamingCollectionEncodingToProtobuf\", \"TranslateStreamCloseWebsocketRequests\", \"UpgradeStatus\", \"UserNamespacesPodSecurityStandards\", \"UserNamespacesSupport\", \"VSphereMultiNetworks\", \"VolumeAttributesClass\", \"VolumeGroupSnapshot\"}}","type":"Normal"},{"namespace":"openshift-machine-config-operator","lastTimestamp":"2025-12-04T11:45:11Z","reason":"SecretCreated","message":"Created Secret/master-user-data-managed -n openshift-machine-api because it was missing","type":"Normal"},{"namespace":"openshift-machine-config-operator","lastTimestamp":"2025-12-04T11:45:11Z","reason":"RoleCreated","message":"Created Role.rbac.authorization.k8s.io/mcd-prometheus-k8s -n openshift-machine-config-operator because it was missing","type":"Normal"},{"namespace":"openshift-machine-config-operator","lastTimestamp":"2025-12-04T11:45:11Z","reason":"SecretCreated","message":"Created Secret/worker-user-data-managed -n openshift-machine-api because it was missing","type":"Normal"},{"namespace":"openshift-machine-config-operator","lastTimestamp":"2025-12-04T11:45:11Z","reason":"RoleCreated","message":"Created Role.rbac.authorization.k8s.io/machine-config-daemon -n openshift-machine-config-operator because it was missing","type":"Normal"},{"namespace":"openshift-machine-config-operator","lastTimestamp":"2025-12-04T11:45:11Z","reason":"ClusterRoleCreated","message":"Created ClusterRole.rbac.authorization.k8s.io/machine-config-daemon because it was missing","type":"Normal"},{"namespace":"openshift-machine-config-operator","lastTimestamp":"2025-12-04T11:45:11Z","reason":"ClusterRoleCreated","message":"Created ClusterRole.rbac.authorization.k8s.io/machine-config-daemon-events because it was missing","type":"Normal"},{"namespace":"openshift-machine-config-operator","lastTimestamp":"2025-12-04T11:45:24Z","reason":"OperatorDegraded: MachineConfigDaemonFailed","message":"Unable to apply 4.18.29: failed to apply machine config daemon manifests: Internal error occurred: admission plugin \"authorization.openshift.io/RestrictSubjectBindings\" failed to complete validation in 13s","type":"Warning"},{"namespace":"openshift-machine-config-operator","lastTimestamp":"2025-12-04T11:45:24Z","reason":"RoleBindingCreateFailed","message":"Failed to create RoleBinding.rbac.authorization.k8s.io/machine-config-daemon-events -n default: Internal error occurred: admission plugin \"authorization.openshift.io/RestrictSubjectBindings\" failed to complete validation in 13s","type":"Warning"},{"namespace":"openshift-machine-config-operator","lastTimestamp":"2025-12-04T11:45:52Z","reason":"OperatorVersionChanged","message":"clusteroperator/machine-config started a version change from [] to [{operator 4.18.29} {operator-image quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b472823604757237c2d16bd6f6221f4cf562aa3b05942c7f602e1e8b2e55a7c6}]","type":"Normal"},{"namespace":"openshift-machine-config-operator","lastTimestamp":"2025-12-04T11:45:53Z","reason":"RoleBindingCreated","message":"Created RoleBinding.rbac.authorization.k8s.io/machine-config-daemon-events -n openshift-machine-config-operator because it was missing","type":"Normal"},{"namespace":"openshift-machine-config-operator","lastTimestamp":"2025-12-04T11:45:53Z","reason":"RoleBindingCreated","message":"Created RoleBinding.rbac.authorization.k8s.io/machine-config-daemon-events -n default because it was missing","type":"Normal"},{"namespace":"openshift-machine-config-operator","lastTimestamp":"2025-12-04T11:45:54Z","reason":"ConfigMapUpdated","message":"Updated ConfigMap/kube-rbac-proxy -n openshift-machine-config-operator:\ncause by changes in data.config-file.yaml","type":"Normal"},{"namespace":"openshift-machine-config-operator","lastTimestamp":"2025-12-04T11:45:54Z","reason":"ValidatingAdmissionPolicyCreated","message":"Created ValidatingAdmissionPolicy.admissionregistration.k8s.io/mcn-guards because it was missing","type":"Normal"},{"namespace":"openshift-machine-config-operator","lastTimestamp":"2025-12-04T11:45:54Z","reason":"ValidatingAdmissionPolicyBindingCreated","message":"Created ValidatingAdmissionPolicyBinding.admissionregistration.k8s.io/mcn-guards-binding because it was missing","type":"Normal"},{"namespace":"openshift-machine-config-operator","lastTimestamp":"2025-12-04T11:45:54Z","reason":"RoleBindingCreated","message":"Created RoleBinding.rbac.authorization.k8s.io/mcd-prometheus-k8s -n openshift-machine-config-operator because it was missing","type":"Normal"},{"namespace":"openshift-machine-config-operator","lastTimestamp":"2025-12-04T11:45:54Z","reason":"RoleBindingCreated","message":"Created RoleBinding.rbac.authorization.k8s.io/machine-config-daemon -n openshift-machine-config-operator because it was missing","type":"Normal"},{"namespace":"openshift-machine-config-operator","lastTimestamp":"2025-12-04T11:45:54Z","reason":"ClusterRoleBindingCreated","message":"Created ClusterRoleBinding.rbac.authorization.k8s.io/machine-config-daemon because it was missing","type":"Normal"},{"namespace":"openshift-machine-config-operator","lastTimestamp":"2025-12-04T11:45:54Z","reason":"ServiceAccountCreated","message":"Created ServiceAccount/machine-config-daemon -n openshift-machine-config-operator because it was missing","type":"Normal"},{"namespace":"openshift-machine-config-operator","lastTimestamp":"2025-12-04T11:53:33Z","reason":"Created","message":"Created container: machine-config-operator","type":"Normal"},{"namespace":"openshift-machine-config-operator","lastTimestamp":"2025-12-04T11:53:33Z","reason":"Started","message":"Started container machine-config-operator","type":"Normal"},{"namespace":"openshift-machine-config-operator","lastTimestamp":"2025-12-04T11:53:33Z","reason":"Pulled","message":"Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b472823604757237c2d16bd6f6221f4cf562aa3b05942c7f602e1e8b2e55a7c6\" already present on machine","type":"Normal"},{"namespace":"openshift-machine-config-operator","lastTimestamp":"2025-12-04T11:59:08Z","reason":"FeatureGatesInitialized","message":"FeatureGates updated to featuregates.Features{Enabled:[]v1.FeatureGateName{\"AWSEFSDriverVolumeMetrics\", \"AdminNetworkPolicy\", \"AlibabaPlatform\", \"AzureWorkloadIdentity\", \"BareMetalLoadBalancer\", \"BuildCSIVolumes\", \"ChunkSizeMiB\", \"CloudDualStackNodeIPs\", \"DisableKubeletCloudCredentialProviders\", \"GCPLabelsTags\", \"HardwareSpeed\", \"IngressControllerLBSubnetsAWS\", \"KMSv1\", \"ManagedBootImages\", \"ManagedBootImagesAWS\", \"MultiArchInstallAWS\", \"MultiArchInstallGCP\", \"NetworkDiagnosticsConfig\", \"NetworkLiveMigration\", \"NetworkSegmentation\", \"NewOLM\", \"NodeDisruptionPolicy\", \"OnClusterBuild\", \"PersistentIPsForVirtualization\", \"PrivateHostedZoneAWS\", \"SetEIPForNLBIngressController\", \"VSphereControlPlaneMachineSet\", \"VSphereDriverConfiguration\", \"VSphereMultiVCenters\", \"VSphereStaticIPs\", \"ValidatingAdmissionPolicy\"}, Disabled:[]v1.FeatureGateName{\"AWSClusterHostedDNS\", \"AdditionalRoutingCapabilities\", \"AutomatedEtcdBackup\", \"BootcNodeManagement\", \"CSIDriverSharedResource\", \"ClusterAPIInstall\", \"ClusterAPIInstallIBMCloud\", \"ClusterMonitoringConfig\", \"ConsolePluginContentSecurityPolicy\", \"DNSNameResolver\", \"DynamicResourceAllocation\", \"EtcdBackendQuota\", \"EventedPLEG\", \"Example\", \"ExternalOIDC\", \"ExternalOIDCWithUIDAndExtraClaimMappings\", \"GCPClusterHostedDNS\", \"GatewayAPI\", \"ImageStreamImportMode\", \"IngressControllerDynamicConfigurationManager\", \"InsightsConfig\", \"InsightsConfigAPI\", \"InsightsOnDemandDataGather\", \"InsightsRuntimeExtractor\", \"MachineAPIMigration\", \"MachineAPIOperatorDisableMachineHealthCheckController\", \"MachineAPIProviderOpenStack\", \"MachineConfigNodes\", \"MaxUnavailableStatefulSet\", \"MetricsCollectionProfiles\", \"MinimumKubeletVersion\", \"MixedCPUsAllocation\", \"MultiArchInstallAzure\", \"NodeSwap\", \"NutanixMultiSubnets\", \"OVNObservability\", \"OpenShiftPodSecurityAdmission\", \"PinnedImages\", \"PlatformOperators\", \"ProcMountType\", \"RouteAdvertisements\", \"RouteExternalCertificate\", \"ServiceAccountTokenNodeBinding\", \"SignatureStores\", \"SigstoreImageVerification\", \"StreamingCollectionEncodingToJSON\", \"StreamingCollectionEncodingToProtobuf\", \"TranslateStreamCloseWebsocketRequests\", \"UpgradeStatus\", \"UserNamespacesPodSecurityStandards\", \"UserNamespacesSupport\", \"VSphereMultiNetworks\", \"VolumeAttributesClass\", \"VolumeGroupSnapshot\"}}","type":"Normal"},{"namespace":"openshift-machine-config-operator","lastTimestamp":"2025-12-04T11:59:10Z","reason":"ConfigMapUpdated","message":"Updated ConfigMap/kube-rbac-proxy -n openshift-machine-config-operator:\ncause by changes in data.config-file.yaml","type":"Normal"},{"namespace":"openshift-machine-config-operator","lastTimestamp":"2025-12-04T11:59:27Z","reason":"FailedMount","message":"MountVolume.SetUp failed for volume \"proxy-tls\" : failed to sync secret cache: timed out waiting for the condition","type":"Warning"},{"namespace":"openshift-machine-config-operator","lastTimestamp":"2025-12-04T11:59:27Z","reason":"FailedMount","message":"MountVolume.SetUp failed for volume \"images\" : failed to sync configmap cache: timed out waiting for the condition","type":"Warning"},{"namespace":"openshift-machine-config-operator","lastTimestamp":"2025-12-04T11:59:27Z","reason":"FailedMount","message":"MountVolume.SetUp failed for volume \"auth-proxy-config\" : failed to sync configmap cache: timed out waiting for the condition","type":"Warning"}]}events/openshift-kube-controller-manager-operator.json0000640000000000000000000015022615114274105022012 0ustar0000000000000000{"items":[{"namespace":"openshift-kube-controller-manager-operator","lastTimestamp":"2025-12-04T11:31:29Z","reason":"ScalingReplicaSet","message":"Scaled up replica set kube-controller-manager-operator-848f645654 to 1","type":"Normal"},{"namespace":"openshift-kube-controller-manager-operator","lastTimestamp":"2025-12-04T11:31:40Z","reason":"FailedCreate","message":"Error creating: pods \"kube-controller-manager-operator-848f645654-\" is forbidden: autoscaling.openshift.io/ManagementCPUsOverride the cluster does not have any nodes","type":"Warning"},{"namespace":"openshift-kube-controller-manager-operator","lastTimestamp":"2025-12-04T11:35:47Z","reason":"FailedCreate","message":"Error creating: pods \"kube-controller-manager-operator-848f645654-\" is forbidden: autoscaling.openshift.io/ManagementCPUsOverride the cluster does not have any nodes","type":"Warning"},{"namespace":"openshift-kube-controller-manager-operator","lastTimestamp":"2025-12-04T11:35:49Z","reason":"FailedScheduling","message":"0/1 nodes are available: 1 node(s) had untolerated taint {node.kubernetes.io/not-ready: }. preemption: 0/1 nodes are available: 1 Preemption is not helpful for scheduling.","type":"Warning"},{"namespace":"openshift-kube-controller-manager-operator","lastTimestamp":"2025-12-04T11:35:49Z","reason":"SuccessfulCreate","message":"Created pod: kube-controller-manager-operator-848f645654-7hmhg","type":"Normal"},{"namespace":"openshift-kube-controller-manager-operator","lastTimestamp":"2025-12-04T11:37:30Z","reason":"Scheduled","message":"Successfully assigned openshift-kube-controller-manager-operator/kube-controller-manager-operator-848f645654-7hmhg to master-0","type":"Normal"},{"namespace":"openshift-kube-controller-manager-operator","lastTimestamp":"2025-12-04T11:37:33Z","reason":"AddedInterface","message":"Add eth0 [10.128.0.16/23] from ovn-kubernetes","type":"Normal"},{"namespace":"openshift-kube-controller-manager-operator","lastTimestamp":"2025-12-04T11:37:33Z","reason":"Pulling","message":"Pulling image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a824e468cf8dd61d347e35b2ee5bc2f815666957647098e21a1bb56ff613e5b9\"","type":"Normal"},{"namespace":"openshift-kube-controller-manager-operator","lastTimestamp":"2025-12-04T11:37:42Z","reason":"Pulled","message":"Successfully pulled image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a824e468cf8dd61d347e35b2ee5bc2f815666957647098e21a1bb56ff613e5b9\" in 8.753s (8.753s including waiting). Image size: 503340749 bytes.","type":"Normal"},{"namespace":"openshift-kube-controller-manager-operator","lastTimestamp":"2025-12-04T11:37:42Z","reason":"Created","message":"Created container: kube-controller-manager-operator","type":"Normal"},{"namespace":"openshift-kube-controller-manager-operator","lastTimestamp":"2025-12-04T11:37:42Z","reason":"Started","message":"Started container kube-controller-manager-operator","type":"Normal"},{"namespace":"openshift-kube-controller-manager-operator","lastTimestamp":"2025-12-04T11:37:44Z","reason":"LeaderElection","message":"kube-controller-manager-operator-848f645654-7hmhg_de738ab7-0991-47f6-ae87-77abd1e92b90 became leader","type":"Normal"},{"namespace":"openshift-kube-controller-manager-operator","lastTimestamp":"2025-12-04T11:37:44Z","reason":"FeatureGatesInitialized","message":"FeatureGates updated to featuregates.Features{Enabled:[]v1.FeatureGateName{\"AWSEFSDriverVolumeMetrics\", \"AdminNetworkPolicy\", \"AlibabaPlatform\", \"AzureWorkloadIdentity\", \"BareMetalLoadBalancer\", \"BuildCSIVolumes\", \"ChunkSizeMiB\", \"CloudDualStackNodeIPs\", \"DisableKubeletCloudCredentialProviders\", \"GCPLabelsTags\", \"HardwareSpeed\", \"IngressControllerLBSubnetsAWS\", \"KMSv1\", \"ManagedBootImages\", \"ManagedBootImagesAWS\", \"MultiArchInstallAWS\", \"MultiArchInstallGCP\", \"NetworkDiagnosticsConfig\", \"NetworkLiveMigration\", \"NetworkSegmentation\", \"NewOLM\", \"NodeDisruptionPolicy\", \"OnClusterBuild\", \"PersistentIPsForVirtualization\", \"PrivateHostedZoneAWS\", \"SetEIPForNLBIngressController\", \"VSphereControlPlaneMachineSet\", \"VSphereDriverConfiguration\", \"VSphereMultiVCenters\", \"VSphereStaticIPs\", \"ValidatingAdmissionPolicy\"}, Disabled:[]v1.FeatureGateName{\"AWSClusterHostedDNS\", \"AdditionalRoutingCapabilities\", \"AutomatedEtcdBackup\", \"BootcNodeManagement\", \"CSIDriverSharedResource\", \"ClusterAPIInstall\", \"ClusterAPIInstallIBMCloud\", \"ClusterMonitoringConfig\", \"ConsolePluginContentSecurityPolicy\", \"DNSNameResolver\", \"DynamicResourceAllocation\", \"EtcdBackendQuota\", \"EventedPLEG\", \"Example\", \"ExternalOIDC\", \"ExternalOIDCWithUIDAndExtraClaimMappings\", \"GCPClusterHostedDNS\", \"GatewayAPI\", \"ImageStreamImportMode\", \"IngressControllerDynamicConfigurationManager\", \"InsightsConfig\", \"InsightsConfigAPI\", \"InsightsOnDemandDataGather\", \"InsightsRuntimeExtractor\", \"MachineAPIMigration\", \"MachineAPIOperatorDisableMachineHealthCheckController\", \"MachineAPIProviderOpenStack\", \"MachineConfigNodes\", \"MaxUnavailableStatefulSet\", \"MetricsCollectionProfiles\", \"MinimumKubeletVersion\", \"MixedCPUsAllocation\", \"MultiArchInstallAzure\", \"NodeSwap\", \"NutanixMultiSubnets\", \"OVNObservability\", \"OpenShiftPodSecurityAdmission\", \"PinnedImages\", \"PlatformOperators\", \"ProcMountType\", \"RouteAdvertisements\", \"RouteExternalCertificate\", \"ServiceAccountTokenNodeBinding\", \"SignatureStores\", \"SigstoreImageVerification\", \"StreamingCollectionEncodingToJSON\", \"StreamingCollectionEncodingToProtobuf\", \"TranslateStreamCloseWebsocketRequests\", \"UpgradeStatus\", \"UserNamespacesPodSecurityStandards\", \"UserNamespacesSupport\", \"VSphereMultiNetworks\", \"VolumeAttributesClass\", \"VolumeGroupSnapshot\"}}","type":"Normal"},{"namespace":"openshift-kube-controller-manager-operator","lastTimestamp":"2025-12-04T11:37:44Z","reason":"OperatorVersionChanged","message":"clusteroperator/kube-controller-manager version \"raw-internal\" changed from \"\" to \"4.18.29\"","type":"Normal"},{"namespace":"openshift-kube-controller-manager-operator","lastTimestamp":"2025-12-04T11:37:44Z","reason":"OperatorStatusChanged","message":"Status for clusteroperator/kube-controller-manager changed: Degraded set to Unknown (\"\"),Progressing set to Unknown (\"\"),Available set to Unknown (\"\"),Upgradeable set to Unknown (\"\"),EvaluationConditionsDetected set to Unknown (\"\"),status.relatedObjects changed from [{\"operator.openshift.io\" \"kubecontrollermanagers\" \"\" \"cluster\"} {\"\" \"namespaces\" \"\" \"openshift-config\"} {\"\" \"namespaces\" \"\" \"openshift-config-managed\"} {\"\" \"namespaces\" \"\" \"openshift-kube-controller-manager\"} {\"\" \"namespaces\" \"\" \"openshift-kube-controller-manager-operator\"} {\"\" \"namespaces\" \"\" \"kube-system\"} {\"\" \"nodes\" \"\" \"\"} {\"certificates.k8s.io\" \"certificatesigningrequests\" \"\" \"\"}] to [{\"operator.openshift.io\" \"kubecontrollermanagers\" \"\" \"cluster\"} {\"\" \"namespaces\" \"\" \"openshift-config\"} {\"\" \"namespaces\" \"\" \"openshift-config-managed\"} {\"\" \"namespaces\" \"\" \"openshift-kube-controller-manager\"} {\"\" \"namespaces\" \"\" \"openshift-kube-controller-manager-operator\"} {\"\" \"namespaces\" \"\" \"kube-system\"} {\"certificates.k8s.io\" \"certificatesigningrequests\" \"\" \"\"} {\"\" \"nodes\" \"\" \"\"} {\"config.openshift.io\" \"nodes\" \"\" \"cluster\"}],status.versions changed from [] to [{\"raw-internal\" \"4.18.29\"}]","type":"Normal"},{"namespace":"openshift-kube-controller-manager-operator","lastTimestamp":"2025-12-04T11:37:44Z","reason":"OperatorStatusChanged","message":"Status for clusteroperator/kube-controller-manager changed: Progressing changed from Unknown to False (\"All is well\")","type":"Normal"},{"namespace":"openshift-kube-controller-manager-operator","lastTimestamp":"2025-12-04T11:37:45Z","reason":"OperatorStatusChanged","message":"Status for clusteroperator/kube-controller-manager changed: Degraded changed from Unknown to False (\"All is well\"),Upgradeable changed from Unknown to True (\"All is well\")","type":"Normal"},{"namespace":"openshift-kube-controller-manager-operator","lastTimestamp":"2025-12-04T11:37:45Z","reason":"CABundleUpdateRequired","message":"\"csr-controller-signer-ca\" in \"openshift-kube-controller-manager-operator\" requires a new cert: configmap doesn't exist","type":"Normal"},{"namespace":"openshift-kube-controller-manager-operator","lastTimestamp":"2025-12-04T11:37:45Z","reason":"MasterNodeObserved","message":"Observed new master node master-0","type":"Normal"},{"namespace":"openshift-kube-controller-manager-operator","lastTimestamp":"2025-12-04T11:37:45Z","reason":"MasterNodesReadyChanged","message":"All master nodes are ready","type":"Normal"},{"namespace":"openshift-kube-controller-manager-operator","lastTimestamp":"2025-12-04T11:37:45Z","reason":"OperatorStatusChanged","message":"Status for clusteroperator/kube-controller-manager changed: Degraded message changed from \"All is well\" to \"NodeControllerDegraded: All master nodes are ready\"","type":"Normal"},{"namespace":"openshift-kube-controller-manager-operator","lastTimestamp":"2025-12-04T11:37:45Z","reason":"ObserveFeatureFlagsUpdated","message":"Updated extendedArguments.feature-gates to AWSEFSDriverVolumeMetrics=true,AdminNetworkPolicy=true,AlibabaPlatform=true,AzureWorkloadIdentity=true,BareMetalLoadBalancer=true,BuildCSIVolumes=true,ChunkSizeMiB=true,CloudDualStackNodeIPs=true,DisableKubeletCloudCredentialProviders=true,GCPLabelsTags=true,HardwareSpeed=true,IngressControllerLBSubnetsAWS=true,KMSv1=true,ManagedBootImages=true,ManagedBootImagesAWS=true,MultiArchInstallAWS=true,MultiArchInstallGCP=true,NetworkDiagnosticsConfig=true,NetworkLiveMigration=true,NetworkSegmentation=true,NewOLM=true,NodeDisruptionPolicy=true,OnClusterBuild=true,PersistentIPsForVirtualization=true,PrivateHostedZoneAWS=true,SetEIPForNLBIngressController=true,VSphereControlPlaneMachineSet=true,VSphereDriverConfiguration=true,VSphereMultiVCenters=true,VSphereStaticIPs=true,ValidatingAdmissionPolicy=true,AWSClusterHostedDNS=false,AdditionalRoutingCapabilities=false,AutomatedEtcdBackup=false,BootcNodeManagement=false,CSIDriverSharedResource=false,ClusterAPIInstall=false,ClusterAPIInstallIBMCloud=false,ClusterMonitoringConfig=false,ConsolePluginContentSecurityPolicy=false,DNSNameResolver=false,DynamicResourceAllocation=false,EtcdBackendQuota=false,EventedPLEG=false,Example=false,ExternalOIDC=false,ExternalOIDCWithUIDAndExtraClaimMappings=false,GCPClusterHostedDNS=false,GatewayAPI=false,ImageStreamImportMode=false,IngressControllerDynamicConfigurationManager=false,InsightsConfig=false,InsightsConfigAPI=false,InsightsOnDemandDataGather=false,InsightsRuntimeExtractor=false,MachineAPIMigration=false,MachineAPIOperatorDisableMachineHealthCheckController=false,MachineAPIProviderOpenStack=false,MachineConfigNodes=false,MaxUnavailableStatefulSet=false,MetricsCollectionProfiles=false,MinimumKubeletVersion=false,MixedCPUsAllocation=false,MultiArchInstallAzure=false,NodeSwap=false,NutanixMultiSubnets=false,OVNObservability=false,OpenShiftPodSecurityAdmission=false,PinnedImages=false,PlatformOperators=false,ProcMountType=false,RouteAdvertisements=false,RouteExternalCertificate=false,ServiceAccountTokenNodeBinding=false,SignatureStores=false,SigstoreImageVerification=false,StreamingCollectionEncodingToJSON=false,StreamingCollectionEncodingToProtobuf=false,TranslateStreamCloseWebsocketRequests=false,UpgradeStatus=false,UserNamespacesPodSecurityStandards=false,UserNamespacesSupport=false,VSphereMultiNetworks=false,VolumeAttributesClass=false,VolumeGroupSnapshot=false","type":"Normal"},{"namespace":"openshift-kube-controller-manager-operator","lastTimestamp":"2025-12-04T11:37:45Z","reason":"ObserveFeatureFlagsUpdated","message":"Updated featureGates to AWSEFSDriverVolumeMetrics=true,AdminNetworkPolicy=true,AlibabaPlatform=true,AzureWorkloadIdentity=true,BareMetalLoadBalancer=true,BuildCSIVolumes=true,ChunkSizeMiB=true,CloudDualStackNodeIPs=true,DisableKubeletCloudCredentialProviders=true,GCPLabelsTags=true,HardwareSpeed=true,IngressControllerLBSubnetsAWS=true,KMSv1=true,ManagedBootImages=true,ManagedBootImagesAWS=true,MultiArchInstallAWS=true,MultiArchInstallGCP=true,NetworkDiagnosticsConfig=true,NetworkLiveMigration=true,NetworkSegmentation=true,NewOLM=true,NodeDisruptionPolicy=true,OnClusterBuild=true,PersistentIPsForVirtualization=true,PrivateHostedZoneAWS=true,SetEIPForNLBIngressController=true,VSphereControlPlaneMachineSet=true,VSphereDriverConfiguration=true,VSphereMultiVCenters=true,VSphereStaticIPs=true,ValidatingAdmissionPolicy=true,AWSClusterHostedDNS=false,AdditionalRoutingCapabilities=false,AutomatedEtcdBackup=false,BootcNodeManagement=false,CSIDriverSharedResource=false,ClusterAPIInstall=false,ClusterAPIInstallIBMCloud=false,ClusterMonitoringConfig=false,ConsolePluginContentSecurityPolicy=false,DNSNameResolver=false,DynamicResourceAllocation=false,EtcdBackendQuota=false,EventedPLEG=false,Example=false,ExternalOIDC=false,ExternalOIDCWithUIDAndExtraClaimMappings=false,GCPClusterHostedDNS=false,GatewayAPI=false,ImageStreamImportMode=false,IngressControllerDynamicConfigurationManager=false,InsightsConfig=false,InsightsConfigAPI=false,InsightsOnDemandDataGather=false,InsightsRuntimeExtractor=false,MachineAPIMigration=false,MachineAPIOperatorDisableMachineHealthCheckController=false,MachineAPIProviderOpenStack=false,MachineConfigNodes=false,MaxUnavailableStatefulSet=false,MetricsCollectionProfiles=false,MinimumKubeletVersion=false,MixedCPUsAllocation=false,MultiArchInstallAzure=false,NodeSwap=false,NutanixMultiSubnets=false,OVNObservability=false,OpenShiftPodSecurityAdmission=false,PinnedImages=false,PlatformOperators=false,ProcMountType=false,RouteAdvertisements=false,RouteExternalCertificate=false,ServiceAccountTokenNodeBinding=false,SignatureStores=false,SigstoreImageVerification=false,StreamingCollectionEncodingToJSON=false,StreamingCollectionEncodingToProtobuf=false,TranslateStreamCloseWebsocketRequests=false,UpgradeStatus=false,UserNamespacesPodSecurityStandards=false,UserNamespacesSupport=false,VSphereMultiNetworks=false,VolumeAttributesClass=false,VolumeGroupSnapshot=false","type":"Normal"},{"namespace":"openshift-kube-controller-manager-operator","lastTimestamp":"2025-12-04T11:37:45Z","reason":"ObserveTLSSecurityProfile","message":"minTLSVersion changed to VersionTLS12","type":"Normal"},{"namespace":"openshift-kube-controller-manager-operator","lastTimestamp":"2025-12-04T11:37:45Z","reason":"ObserveTLSSecurityProfile","message":"cipherSuites changed to [\"TLS_AES_128_GCM_SHA256\" \"TLS_AES_256_GCM_SHA384\" \"TLS_CHACHA20_POLY1305_SHA256\" \"TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256\" \"TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256\" \"TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384\" \"TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384\" \"TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305_SHA256\" \"TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305_SHA256\"]","type":"Normal"},{"namespace":"openshift-kube-controller-manager-operator","lastTimestamp":"2025-12-04T11:37:45Z","reason":"ObservedConfigChanged","message":"Writing updated observed config: map[string]any{\n+ \t\"extendedArguments\": map[string]any{\n+ \t\t\"cluster-cidr\": []any{string(\"10.128.0.0/16\")},\n+ \t\t\"cluster-name\": []any{string(\"sno-d7v9r\")},\n+ \t\t\"feature-gates\": []any{\n+ \t\t\tstring(\"AWSEFSDriverVolumeMetrics=true\"), string(\"AdminNetworkPolicy=true\"),\n+ \t\t\tstring(\"AlibabaPlatform=true\"), string(\"AzureWorkloadIdentity=true\"),\n+ \t\t\tstring(\"BareMetalLoadBalancer=true\"), string(\"BuildCSIVolumes=true\"),\n+ \t\t\tstring(\"ChunkSizeMiB=true\"), string(\"CloudDualStackNodeIPs=true\"), ...,\n+ \t\t},\n+ \t\t\"service-cluster-ip-range\": []any{string(\"172.30.0.0/16\")},\n+ \t},\n+ \t\"featureGates\": []any{\n+ \t\tstring(\"AWSEFSDriverVolumeMetrics=true\"), string(\"AdminNetworkPolicy=true\"),\n+ \t\tstring(\"AlibabaPlatform=true\"), string(\"AzureWorkloadIdentity=true\"),\n+ \t\tstring(\"BareMetalLoadBalancer=true\"), string(\"BuildCSIVolumes=true\"),\n+ \t\tstring(\"ChunkSizeMiB=true\"), string(\"CloudDualStackNodeIPs=true\"),\n+ \t\tstring(\"DisableKubeletCloudCredentialProviders=true\"),\n+ \t\tstring(\"GCPLabelsTags=true\"), string(\"HardwareSpeed=true\"),\n+ \t\tstring(\"IngressControllerLBSubnetsAWS=true\"), string(\"KMSv1=true\"),\n+ \t\tstring(\"ManagedBootImages=true\"), string(\"ManagedBootImagesAWS=true\"),\n+ \t\tstring(\"MultiArchInstallAWS=true\"), ...,\n+ \t},\n+ \t\"servingInfo\": map[string]any{\n+ \t\t\"cipherSuites\": []any{\n+ \t\t\tstring(\"TLS_AES_128_GCM_SHA256\"), string(\"TLS_AES_256_GCM_SHA384\"),\n+ \t\t\tstring(\"TLS_CHACHA20_POLY1305_SHA256\"),\n+ \t\t\tstring(\"TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256\"),\n+ \t\t\tstring(\"TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256\"),\n+ \t\t\tstring(\"TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384\"),\n+ \t\t\tstring(\"TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384\"),\n+ \t\t\tstring(\"TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305_SHA256\"), ...,\n+ \t\t},\n+ \t\t\"minTLSVersion\": string(\"VersionTLS12\"),\n+ \t},\n }\n","type":"Normal"},{"namespace":"openshift-kube-controller-manager-operator","lastTimestamp":"2025-12-04T11:37:46Z","reason":"OperatorStatusChanged","message":"Status for clusteroperator/kube-controller-manager changed: Degraded message changed from \"NodeControllerDegraded: All master nodes are ready\" to \"NodeControllerDegraded: All master nodes are ready\\nRevisionControllerDegraded: configmap \\\"kube-controller-manager-pod\\\" not found\"","type":"Normal"},{"namespace":"openshift-kube-controller-manager-operator","lastTimestamp":"2025-12-04T11:37:46Z","reason":"OperatorStatusChanged","message":"Status for clusteroperator/kube-controller-manager changed: Degraded message changed from \"NodeControllerDegraded: All master nodes are ready\\nRevisionControllerDegraded: configmap \\\"kube-controller-manager-pod\\\" not found\" to \"NodeControllerDegraded: All master nodes are ready\\nRevisionControllerDegraded: configmap \\\"kube-controller-manager-pod\\\" not found\\nInstallerControllerDegraded: missing required resources: [configmaps: cluster-policy-controller-config-0,config-0,controller-manager-kubeconfig-0,kube-controller-cert-syncer-kubeconfig-0,kube-controller-manager-pod-0,recycler-config-0,service-ca-0,serviceaccount-ca-0, secrets: localhost-recovery-client-token-0,service-account-private-key-0]\",Progressing message changed from \"All is well\" to \"NodeInstallerProgressing: 1 node is at revision 0\",Available changed from Unknown to False (\"StaticPodsAvailable: 0 nodes are active; 1 node is at revision 0\")","type":"Normal"},{"namespace":"openshift-kube-controller-manager-operator","lastTimestamp":"2025-12-04T11:37:47Z","reason":"ConfigMapCreated","message":"Created ConfigMap/csr-controller-signer-ca -n openshift-kube-controller-manager-operator because it was missing","type":"Normal"},{"namespace":"openshift-kube-controller-manager-operator","lastTimestamp":"2025-12-04T11:37:47Z","reason":"TargetUpdateRequired","message":"\"csr-signer\" in \"openshift-kube-controller-manager-operator\" requires a new target cert/key pair: secret doesn't exist","type":"Normal"},{"namespace":"openshift-kube-controller-manager-operator","lastTimestamp":"2025-12-04T11:37:48Z","reason":"ConfigMapCreated","message":"Created ConfigMap/aggregator-client-ca -n openshift-kube-controller-manager because it was missing","type":"Normal"},{"namespace":"openshift-kube-controller-manager-operator","lastTimestamp":"2025-12-04T11:37:48Z","reason":"SecretCreated","message":"Created Secret/csr-signer -n openshift-kube-controller-manager-operator because it was missing","type":"Normal"},{"namespace":"openshift-kube-controller-manager-operator","lastTimestamp":"2025-12-04T11:37:49Z","reason":"ServiceAccountCreated","message":"Created ServiceAccount/installer-sa -n openshift-kube-controller-manager because it was missing","type":"Normal"},{"namespace":"openshift-kube-controller-manager-operator","lastTimestamp":"2025-12-04T11:37:49Z","reason":"ClusterRoleBindingCreated","message":"Created ClusterRoleBinding.rbac.authorization.k8s.io/system:openshift:operator:openshift-kube-controller-manager-installer because it was missing","type":"Normal"},{"namespace":"openshift-kube-controller-manager-operator","lastTimestamp":"2025-12-04T11:37:49Z","reason":"TargetConfigDeleted","message":"Deleted target configmap openshift-config-managed/csr-controller-ca because source config does not exist","type":"Normal"},{"namespace":"openshift-kube-controller-manager-operator","lastTimestamp":"2025-12-04T11:37:49Z","reason":"NamespaceUpdated","message":"Updated Namespace/openshift-kube-controller-manager because it changed","type":"Normal"},{"namespace":"openshift-kube-controller-manager-operator","lastTimestamp":"2025-12-04T11:37:49Z","reason":"ConfigMapCreated","message":"Created ConfigMap/config -n openshift-kube-controller-manager because it was missing","type":"Normal"},{"namespace":"openshift-kube-controller-manager-operator","lastTimestamp":"2025-12-04T11:37:50Z","reason":"SecretCreated","message":"Created Secret/service-account-private-key -n openshift-kube-controller-manager because it was missing","type":"Normal"},{"namespace":"openshift-kube-controller-manager-operator","lastTimestamp":"2025-12-04T11:37:50Z","reason":"ConfigMapCreated","message":"Created ConfigMap/service-ca -n openshift-kube-controller-manager because it was missing","type":"Normal"},{"namespace":"openshift-kube-controller-manager-operator","lastTimestamp":"2025-12-04T11:37:50Z","reason":"ConfigMapCreated","message":"Created ConfigMap/kube-controller-cert-syncer-kubeconfig -n openshift-kube-controller-manager because it was missing","type":"Normal"},{"namespace":"openshift-kube-controller-manager-operator","lastTimestamp":"2025-12-04T11:37:50Z","reason":"RoleBindingCreated","message":"Created RoleBinding.rbac.authorization.k8s.io/system:openshift:leader-locking-kube-controller-manager -n kube-system because it was missing","type":"Normal"},{"namespace":"openshift-kube-controller-manager-operator","lastTimestamp":"2025-12-04T11:37:50Z","reason":"RoleCreated","message":"Created Role.rbac.authorization.k8s.io/system:openshift:leader-election-lock-cluster-policy-controller -n openshift-kube-controller-manager because it was missing","type":"Normal"},{"namespace":"openshift-kube-controller-manager-operator","lastTimestamp":"2025-12-04T11:37:51Z","reason":"RoleBindingCreated","message":"Created RoleBinding.rbac.authorization.k8s.io/system:openshift:leader-election-lock-cluster-policy-controller -n openshift-kube-controller-manager because it was missing","type":"Normal"},{"namespace":"openshift-kube-controller-manager-operator","lastTimestamp":"2025-12-04T11:37:51Z","reason":"SecretCreated","message":"Created Secret/kube-controller-manager-client-cert-key -n openshift-kube-controller-manager because it was missing","type":"Normal"},{"namespace":"openshift-kube-controller-manager-operator","lastTimestamp":"2025-12-04T11:37:51Z","reason":"ObserveServiceCAConfigMap","message":"observed change in config","type":"Normal"},{"namespace":"openshift-kube-controller-manager-operator","lastTimestamp":"2025-12-04T11:37:51Z","reason":"ObservedConfigChanged","message":"Writing updated observed config: map[string]any{\n \t\"extendedArguments\": map[string]any{\"cluster-cidr\": []any{string(\"10.128.0.0/16\")}, \"cluster-name\": []any{string(\"sno-d7v9r\")}, \"feature-gates\": []any{string(\"AWSEFSDriverVolumeMetrics=true\"), string(\"AdminNetworkPolicy=true\"), string(\"AlibabaPlatform=true\"), string(\"AzureWorkloadIdentity=true\"), ...}, \"service-cluster-ip-range\": []any{string(\"172.30.0.0/16\")}},\n \t\"featureGates\": []any{string(\"AWSEFSDriverVolumeMetrics=true\"), string(\"AdminNetworkPolicy=true\"), string(\"AlibabaPlatform=true\"), string(\"AzureWorkloadIdentity=true\"), ...},\n+ \t\"serviceServingCert\": map[string]any{\n+ \t\t\"certFile\": string(\"/etc/kubernetes/static-pod-resources/configmaps/service-ca/ca-bundle.crt\"),\n+ \t},\n \t\"servingInfo\": map[string]any{\"cipherSuites\": []any{string(\"TLS_AES_128_GCM_SHA256\"), string(\"TLS_AES_256_GCM_SHA384\"), string(\"TLS_CHACHA20_POLY1305_SHA256\"), string(\"TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256\"), ...}, \"minTLSVersion\": string(\"VersionTLS12\")},\n }\n","type":"Normal"},{"namespace":"openshift-kube-controller-manager-operator","lastTimestamp":"2025-12-04T11:37:53Z","reason":"ConfigMapCreated","message":"Created ConfigMap/cluster-policy-controller-config -n openshift-kube-controller-manager because it was missing","type":"Normal"},{"namespace":"openshift-kube-controller-manager-operator","lastTimestamp":"2025-12-04T11:37:54Z","reason":"ServiceCreated","message":"Created Service/kube-controller-manager -n openshift-kube-controller-manager because it was missing","type":"Normal"},{"namespace":"openshift-kube-controller-manager-operator","lastTimestamp":"2025-12-04T11:37:54Z","reason":"ConfigMapCreated","message":"Created ConfigMap/recycler-config -n openshift-kube-controller-manager because it was missing","type":"Normal"},{"namespace":"openshift-kube-controller-manager-operator","lastTimestamp":"2025-12-04T11:37:55Z","reason":"ServiceAccountCreated","message":"Created ServiceAccount/kube-controller-manager-sa -n openshift-kube-controller-manager because it was missing","type":"Normal"},{"namespace":"openshift-kube-controller-manager-operator","lastTimestamp":"2025-12-04T11:37:57Z","reason":"ConfigMapCreated","message":"Created ConfigMap/csr-signer-ca -n openshift-kube-controller-manager-operator because it was missing","type":"Normal"},{"namespace":"openshift-kube-controller-manager-operator","lastTimestamp":"2025-12-04T11:37:57Z","reason":"ServiceAccountCreated","message":"Created ServiceAccount/pv-recycler-controller -n openshift-infra because it was missing","type":"Normal"},{"namespace":"openshift-kube-controller-manager-operator","lastTimestamp":"2025-12-04T11:37:57Z","reason":"ClusterRoleBindingCreated","message":"Created ClusterRoleBinding.rbac.authorization.k8s.io/system:openshift:operator:kube-controller-manager-recovery because it was missing","type":"Normal"},{"namespace":"openshift-kube-controller-manager-operator","lastTimestamp":"2025-12-04T11:37:58Z","reason":"ConfigMapCreated","message":"Created ConfigMap/csr-controller-ca -n openshift-kube-controller-manager-operator because it was missing","type":"Normal"},{"namespace":"openshift-kube-controller-manager-operator","lastTimestamp":"2025-12-04T11:37:58Z","reason":"ServiceAccountCreated","message":"Created ServiceAccount/localhost-recovery-client -n openshift-kube-controller-manager because it was missing","type":"Normal"},{"namespace":"openshift-kube-controller-manager-operator","lastTimestamp":"2025-12-04T11:37:59Z","reason":"ConfigMapCreated","message":"Created ConfigMap/csr-controller-ca -n openshift-config-managed because it was missing","type":"Normal"},{"namespace":"openshift-kube-controller-manager-operator","lastTimestamp":"2025-12-04T11:37:59Z","reason":"SecretCreated","message":"Created Secret/localhost-recovery-client-token -n openshift-kube-controller-manager because it was missing","type":"Normal"},{"namespace":"openshift-kube-controller-manager-operator","lastTimestamp":"2025-12-04T11:37:59Z","reason":"ClusterRoleCreated","message":"Created ClusterRole.rbac.authorization.k8s.io/system:openshift:controller:cluster-csr-approver-controller because it was missing","type":"Normal"},{"namespace":"openshift-kube-controller-manager-operator","lastTimestamp":"2025-12-04T11:37:59Z","reason":"ClusterRoleBindingCreated","message":"Created ClusterRoleBinding.rbac.authorization.k8s.io/system:openshift:controller:cluster-csr-approver-controller because it was missing","type":"Normal"},{"namespace":"openshift-kube-controller-manager-operator","lastTimestamp":"2025-12-04T11:38:00Z","reason":"ConfigMapCreated","message":"Created ConfigMap/client-ca -n openshift-kube-controller-manager because it was missing","type":"Normal"},{"namespace":"openshift-kube-controller-manager-operator","lastTimestamp":"2025-12-04T11:38:01Z","reason":"SecretCreated","message":"Created Secret/csr-signer -n openshift-kube-controller-manager because it was missing","type":"Normal"},{"namespace":"openshift-kube-controller-manager-operator","lastTimestamp":"2025-12-04T11:38:03Z","reason":"ConfigMapCreated","message":"Created ConfigMap/serviceaccount-ca -n openshift-kube-controller-manager because it was missing","type":"Normal"},{"namespace":"openshift-kube-controller-manager-operator","lastTimestamp":"2025-12-04T11:38:05Z","reason":"ConfigMapCreated","message":"Created ConfigMap/controller-manager-kubeconfig -n openshift-kube-controller-manager because it was missing","type":"Normal"},{"namespace":"openshift-kube-controller-manager-operator","lastTimestamp":"2025-12-04T11:38:12Z","reason":"StartingNewRevision","message":"new revision 1 triggered by \"configmap \\\"kube-controller-manager-pod-0\\\" not found\"","type":"Normal"},{"namespace":"openshift-kube-controller-manager-operator","lastTimestamp":"2025-12-04T11:38:12Z","reason":"ConfigMapCreated","message":"Created ConfigMap/kube-controller-manager-pod -n openshift-kube-controller-manager because it was missing","type":"Normal"},{"namespace":"openshift-kube-controller-manager-operator","lastTimestamp":"2025-12-04T11:38:18Z","reason":"ConfigMapCreated","message":"Created ConfigMap/kube-controller-manager-pod-1 -n openshift-kube-controller-manager because it was missing","type":"Normal"},{"namespace":"openshift-kube-controller-manager-operator","lastTimestamp":"2025-12-04T11:38:18Z","reason":"ConfigMapCreated","message":"Created ConfigMap/config-1 -n openshift-kube-controller-manager because it was missing","type":"Normal"},{"namespace":"openshift-kube-controller-manager-operator","lastTimestamp":"2025-12-04T11:38:18Z","reason":"ConfigMapUpdated","message":"Updated ConfigMap/client-ca -n openshift-kube-controller-manager:\ncause by changes in data.ca-bundle.crt","type":"Normal"},{"namespace":"openshift-kube-controller-manager-operator","lastTimestamp":"2025-12-04T11:38:18Z","reason":"ConfigMapCreated","message":"Created ConfigMap/cluster-policy-controller-config-1 -n openshift-kube-controller-manager because it was missing","type":"Normal"},{"namespace":"openshift-kube-controller-manager-operator","lastTimestamp":"2025-12-04T11:38:18Z","reason":"ConfigMapCreated","message":"Created ConfigMap/controller-manager-kubeconfig-1 -n openshift-kube-controller-manager because it was missing","type":"Normal"},{"namespace":"openshift-kube-controller-manager-operator","lastTimestamp":"2025-12-04T11:38:18Z","reason":"ConfigMapCreated","message":"Created ConfigMap/kube-controller-cert-syncer-kubeconfig-1 -n openshift-kube-controller-manager because it was missing","type":"Normal"},{"namespace":"openshift-kube-controller-manager-operator","lastTimestamp":"2025-12-04T11:38:18Z","reason":"ConfigMapCreated","message":"Created ConfigMap/serviceaccount-ca-1 -n openshift-kube-controller-manager because it was missing","type":"Normal"},{"namespace":"openshift-kube-controller-manager-operator","lastTimestamp":"2025-12-04T11:38:18Z","reason":"ConfigMapCreated","message":"Created ConfigMap/service-ca-1 -n openshift-kube-controller-manager because it was missing","type":"Normal"},{"namespace":"openshift-kube-controller-manager-operator","lastTimestamp":"2025-12-04T11:38:19Z","reason":"ConfigMapCreated","message":"Created ConfigMap/recycler-config-1 -n openshift-kube-controller-manager because it was missing","type":"Normal"},{"namespace":"openshift-kube-controller-manager-operator","lastTimestamp":"2025-12-04T11:38:20Z","reason":"SecretCreated","message":"Created Secret/service-account-private-key-1 -n openshift-kube-controller-manager because it was missing","type":"Normal"},{"namespace":"openshift-kube-controller-manager-operator","lastTimestamp":"2025-12-04T11:38:20Z","reason":"ConfigMapUpdated","message":"Updated ConfigMap/config -n openshift-kube-controller-manager:\ncause by changes in data.config.yaml","type":"Normal"},{"namespace":"openshift-kube-controller-manager-operator","lastTimestamp":"2025-12-04T11:38:21Z","reason":"SecretCreated","message":"Created Secret/serving-cert-1 -n openshift-kube-controller-manager because it was missing","type":"Normal"},{"namespace":"openshift-kube-controller-manager-operator","lastTimestamp":"2025-12-04T11:38:22Z","reason":"SecretCreated","message":"Created Secret/localhost-recovery-client-token-1 -n openshift-kube-controller-manager because it was missing","type":"Normal"},{"namespace":"openshift-kube-controller-manager-operator","lastTimestamp":"2025-12-04T11:38:23Z","reason":"ConfigMapUpdated","message":"Updated ConfigMap/cluster-policy-controller-config -n openshift-kube-controller-manager:\ncause by changes in data.config.yaml","type":"Normal"},{"namespace":"openshift-kube-controller-manager-operator","lastTimestamp":"2025-12-04T11:38:23Z","reason":"RequiredInstallerResourcesMissing","message":"configmaps: cluster-policy-controller-config-0,config-0,controller-manager-kubeconfig-0,kube-controller-cert-syncer-kubeconfig-0,kube-controller-manager-pod-0,recycler-config-0,service-ca-0,serviceaccount-ca-0, secrets: localhost-recovery-client-token-0,service-account-private-key-0","type":"Warning"},{"namespace":"openshift-kube-controller-manager-operator","lastTimestamp":"2025-12-04T11:38:23Z","reason":"RevisionTriggered","message":"new revision 1 triggered by \"configmap \\\"kube-controller-manager-pod-0\\\" not found\"","type":"Normal"},{"namespace":"openshift-kube-controller-manager-operator","lastTimestamp":"2025-12-04T11:38:23Z","reason":"StartingNewRevision","message":"new revision 2 triggered by \"required configmap/config has changed\"","type":"Normal"},{"namespace":"openshift-kube-controller-manager-operator","lastTimestamp":"2025-12-04T11:38:23Z","reason":"OperatorStatusChanged","message":"Status for clusteroperator/kube-controller-manager changed: Degraded message changed from \"NodeControllerDegraded: All master nodes are ready\\nRevisionControllerDegraded: configmap \\\"kube-controller-manager-pod\\\" not found\\nInstallerControllerDegraded: missing required resources: [configmaps: cluster-policy-controller-config-0,config-0,controller-manager-kubeconfig-0,kube-controller-cert-syncer-kubeconfig-0,kube-controller-manager-pod-0,recycler-config-0,service-ca-0,serviceaccount-ca-0, secrets: localhost-recovery-client-token-0,service-account-private-key-0]\" to \"NodeControllerDegraded: All master nodes are ready\\nInstallerControllerDegraded: missing required resources: [configmaps: cluster-policy-controller-config-0,config-0,controller-manager-kubeconfig-0,kube-controller-cert-syncer-kubeconfig-0,kube-controller-manager-pod-0,recycler-config-0,service-ca-0,serviceaccount-ca-0, secrets: localhost-recovery-client-token-0,service-account-private-key-0]\"","type":"Normal"},{"namespace":"openshift-kube-controller-manager-operator","lastTimestamp":"2025-12-04T11:38:24Z","reason":"OperatorStatusChanged","message":"Status for clusteroperator/kube-controller-manager changed: Degraded message changed from \"NodeControllerDegraded: All master nodes are ready\\nInstallerControllerDegraded: missing required resources: [configmaps: cluster-policy-controller-config-0,config-0,controller-manager-kubeconfig-0,kube-controller-cert-syncer-kubeconfig-0,kube-controller-manager-pod-0,recycler-config-0,service-ca-0,serviceaccount-ca-0, secrets: localhost-recovery-client-token-0,service-account-private-key-0]\" to \"NodeControllerDegraded: All master nodes are ready\",Progressing changed from False to True (\"NodeInstallerProgressing: 1 node is at revision 0; 0 nodes have achieved new revision 1\"),Available message changed from \"StaticPodsAvailable: 0 nodes are active; 1 node is at revision 0\" to \"StaticPodsAvailable: 0 nodes are active; 1 node is at revision 0; 0 nodes have achieved new revision 1\"","type":"Normal"},{"namespace":"openshift-kube-controller-manager-operator","lastTimestamp":"2025-12-04T11:38:24Z","reason":"NodeTargetRevisionChanged","message":"Updating node \"master-0\" from revision 0 to 1 because node master-0 static pod not found","type":"Normal"},{"namespace":"openshift-kube-controller-manager-operator","lastTimestamp":"2025-12-04T11:38:25Z","reason":"ConfigMapCreated","message":"Created ConfigMap/kube-controller-manager-pod-2 -n openshift-kube-controller-manager because it was missing","type":"Normal"},{"namespace":"openshift-kube-controller-manager-operator","lastTimestamp":"2025-12-04T11:38:27Z","reason":"PodCreated","message":"Created Pod/installer-1-master-0 -n openshift-kube-controller-manager because it was missing","type":"Normal"},{"namespace":"openshift-kube-controller-manager-operator","lastTimestamp":"2025-12-04T11:38:27Z","reason":"ConfigMapCreated","message":"Created ConfigMap/config-2 -n openshift-kube-controller-manager because it was missing","type":"Normal"},{"namespace":"openshift-kube-controller-manager-operator","lastTimestamp":"2025-12-04T11:38:28Z","reason":"ConfigMapCreated","message":"Created ConfigMap/cluster-policy-controller-config-2 -n openshift-kube-controller-manager because it was missing","type":"Normal"},{"namespace":"openshift-kube-controller-manager-operator","lastTimestamp":"2025-12-04T11:38:30Z","reason":"ConfigMapCreated","message":"Created ConfigMap/controller-manager-kubeconfig-2 -n openshift-kube-controller-manager because it was missing","type":"Normal"},{"namespace":"openshift-kube-controller-manager-operator","lastTimestamp":"2025-12-04T11:38:31Z","reason":"ConfigMapCreated","message":"Created ConfigMap/kube-controller-cert-syncer-kubeconfig-2 -n openshift-kube-controller-manager because it was missing","type":"Normal"},{"namespace":"openshift-kube-controller-manager-operator","lastTimestamp":"2025-12-04T11:38:32Z","reason":"ConfigMapUpdated","message":"Updated ConfigMap/serviceaccount-ca -n openshift-kube-controller-manager:\ncause by changes in data.ca-bundle.crt","type":"Normal"},{"namespace":"openshift-kube-controller-manager-operator","lastTimestamp":"2025-12-04T11:38:32Z","reason":"ConfigMapCreated","message":"Created ConfigMap/serviceaccount-ca-2 -n openshift-kube-controller-manager because it was missing","type":"Normal"},{"namespace":"openshift-kube-controller-manager-operator","lastTimestamp":"2025-12-04T11:38:33Z","reason":"ConfigMapCreated","message":"Created ConfigMap/recycler-config-2 -n openshift-kube-controller-manager because it was missing","type":"Normal"},{"namespace":"openshift-kube-controller-manager-operator","lastTimestamp":"2025-12-04T11:38:33Z","reason":"OperatorStatusChanged","message":"Status for clusteroperator/kube-controller-manager changed: Degraded message changed from \"NodeControllerDegraded: All master nodes are ready\" to \"GarbageCollectorDegraded: error fetching rules: Get \\\"https://thanos-querier.openshift-monitoring.svc:9091/api/v1/rules\\\": dial tcp: lookup thanos-querier.openshift-monitoring.svc on 172.30.0.10:53: read udp 10.128.0.16:46654-\u003e172.30.0.10:53: read: connection refused\\nNodeControllerDegraded: All master nodes are ready\"","type":"Normal"},{"namespace":"openshift-kube-controller-manager-operator","lastTimestamp":"2025-12-04T11:38:33Z","reason":"OperatorStatusChanged","message":"Status for clusteroperator/kube-controller-manager changed: Degraded message changed from \"GarbageCollectorDegraded: error fetching rules: Get \\\"https://thanos-querier.openshift-monitoring.svc:9091/api/v1/rules\\\": dial tcp: lookup thanos-querier.openshift-monitoring.svc on 172.30.0.10:53: read udp 10.128.0.16:46654-\u003e172.30.0.10:53: read: connection refused\\nNodeControllerDegraded: All master nodes are ready\" to \"GarbageCollectorDegraded: error fetching rules: Get \\\"https://thanos-querier.openshift-monitoring.svc:9091/api/v1/rules\\\": dial tcp: lookup thanos-querier.openshift-monitoring.svc on 172.30.0.10:53: read udp 10.128.0.16:42373-\u003e172.30.0.10:53: read: connection refused\\nNodeControllerDegraded: All master nodes are ready\"","type":"Normal"},{"namespace":"openshift-kube-controller-manager-operator","lastTimestamp":"2025-12-04T11:38:33Z","reason":"OperatorStatusChanged","message":"Status for clusteroperator/kube-controller-manager changed: Degraded message changed from \"GarbageCollectorDegraded: error fetching rules: Get \\\"https://thanos-querier.openshift-monitoring.svc:9091/api/v1/rules\\\": dial tcp: lookup thanos-querier.openshift-monitoring.svc on 172.30.0.10:53: read udp 10.128.0.16:42373-\u003e172.30.0.10:53: read: connection refused\\nNodeControllerDegraded: All master nodes are ready\" to \"NodeControllerDegraded: All master nodes are ready\"","type":"Normal"},{"namespace":"openshift-kube-controller-manager-operator","lastTimestamp":"2025-12-04T11:38:33Z","reason":"ConfigMapCreated","message":"Created ConfigMap/service-ca-2 -n openshift-kube-controller-manager because it was missing","type":"Normal"},{"namespace":"openshift-kube-controller-manager-operator","lastTimestamp":"2025-12-04T11:38:35Z","reason":"SecretCreated","message":"Created Secret/service-account-private-key-2 -n openshift-kube-controller-manager because it was missing","type":"Normal"},{"namespace":"openshift-kube-controller-manager-operator","lastTimestamp":"2025-12-04T11:38:36Z","reason":"SecretCreated","message":"Created Secret/serving-cert-2 -n openshift-kube-controller-manager because it was missing","type":"Normal"},{"namespace":"openshift-kube-controller-manager-operator","lastTimestamp":"2025-12-04T11:38:38Z","reason":"SecretCreated","message":"Created Secret/localhost-recovery-client-token-2 -n openshift-kube-controller-manager because it was missing","type":"Normal"},{"namespace":"openshift-kube-controller-manager-operator","lastTimestamp":"2025-12-04T11:38:40Z","reason":"RevisionTriggered","message":"new revision 2 triggered by \"required configmap/config has changed\"","type":"Normal"},{"namespace":"openshift-kube-controller-manager-operator","lastTimestamp":"2025-12-04T11:38:40Z","reason":"ConfigMapUpdated","message":"Updated ConfigMap/kube-controller-manager-pod -n openshift-kube-controller-manager:\ncause by changes in data.pod.yaml","type":"Normal"},{"namespace":"openshift-kube-controller-manager-operator","lastTimestamp":"2025-12-04T11:38:41Z","reason":"StartingNewRevision","message":"new revision 3 triggered by \"required configmap/serviceaccount-ca has changed\"","type":"Normal"},{"namespace":"openshift-kube-controller-manager-operator","lastTimestamp":"2025-12-04T11:38:42Z","reason":"ConfigMapCreated","message":"Created ConfigMap/kube-controller-manager-pod-3 -n openshift-kube-controller-manager because it was missing","type":"Normal"},{"namespace":"openshift-kube-controller-manager-operator","lastTimestamp":"2025-12-04T11:38:43Z","reason":"OperatorStatusChanged","message":"Status for clusteroperator/kube-controller-manager changed: Progressing message changed from \"NodeInstallerProgressing: 1 node is at revision 0; 0 nodes have achieved new revision 1\" to \"NodeInstallerProgressing: 1 node is at revision 0; 0 nodes have achieved new revision 2\",Available message changed from \"StaticPodsAvailable: 0 nodes are active; 1 node is at revision 0; 0 nodes have achieved new revision 1\" to \"StaticPodsAvailable: 0 nodes are active; 1 node is at revision 0; 0 nodes have achieved new revision 2\"","type":"Normal"},{"namespace":"openshift-kube-controller-manager-operator","lastTimestamp":"2025-12-04T11:38:44Z","reason":"ConfigMapCreated","message":"Created ConfigMap/config-3 -n openshift-kube-controller-manager because it was missing","type":"Normal"},{"namespace":"openshift-kube-controller-manager-operator","lastTimestamp":"2025-12-04T11:38:45Z","reason":"ConfigMapCreated","message":"Created ConfigMap/cluster-policy-controller-config-3 -n openshift-kube-controller-manager because it was missing","type":"Normal"},{"namespace":"openshift-kube-controller-manager-operator","lastTimestamp":"2025-12-04T11:38:46Z","reason":"PodCreated","message":"Created Pod/installer-2-master-0 -n openshift-kube-controller-manager because it was missing","type":"Normal"},{"namespace":"openshift-kube-controller-manager-operator","lastTimestamp":"2025-12-04T11:38:47Z","reason":"ConfigMapCreated","message":"Created ConfigMap/controller-manager-kubeconfig-3 -n openshift-kube-controller-manager because it was missing","type":"Normal"},{"namespace":"openshift-kube-controller-manager-operator","lastTimestamp":"2025-12-04T11:38:48Z","reason":"ConfigMapCreated","message":"Created ConfigMap/kube-controller-cert-syncer-kubeconfig-3 -n openshift-kube-controller-manager because it was missing","type":"Normal"},{"namespace":"openshift-kube-controller-manager-operator","lastTimestamp":"2025-12-04T11:38:49Z","reason":"SATokenSignerControllerStuck","message":"unexpected addresses: 192.168.32.10","type":"Normal"},{"namespace":"openshift-kube-controller-manager-operator","lastTimestamp":"2025-12-04T11:38:50Z","reason":"ConfigMapCreated","message":"Created ConfigMap/serviceaccount-ca-3 -n openshift-kube-controller-manager because it was missing","type":"Normal"},{"namespace":"openshift-kube-controller-manager-operator","lastTimestamp":"2025-12-04T11:38:51Z","reason":"ConfigMapCreated","message":"Created ConfigMap/service-ca-3 -n openshift-kube-controller-manager because it was missing","type":"Normal"},{"namespace":"openshift-kube-controller-manager-operator","lastTimestamp":"2025-12-04T11:50:10Z","reason":"LeaderElection","message":"kube-controller-manager-operator-848f645654-7hmhg_69ca551c-5a70-48b5-b047-ad00c925bdb9 became leader","type":"Normal"},{"namespace":"openshift-kube-controller-manager-operator","lastTimestamp":"2025-12-04T11:50:10Z","reason":"FeatureGatesInitialized","message":"FeatureGates updated to featuregates.Features{Enabled:[]v1.FeatureGateName{\"AWSEFSDriverVolumeMetrics\", \"AdminNetworkPolicy\", \"AlibabaPlatform\", \"AzureWorkloadIdentity\", \"BareMetalLoadBalancer\", \"BuildCSIVolumes\", \"ChunkSizeMiB\", \"CloudDualStackNodeIPs\", \"DisableKubeletCloudCredentialProviders\", \"GCPLabelsTags\", \"HardwareSpeed\", \"IngressControllerLBSubnetsAWS\", \"KMSv1\", \"ManagedBootImages\", \"ManagedBootImagesAWS\", \"MultiArchInstallAWS\", \"MultiArchInstallGCP\", \"NetworkDiagnosticsConfig\", \"NetworkLiveMigration\", \"NetworkSegmentation\", \"NewOLM\", \"NodeDisruptionPolicy\", \"OnClusterBuild\", \"PersistentIPsForVirtualization\", \"PrivateHostedZoneAWS\", \"SetEIPForNLBIngressController\", \"VSphereControlPlaneMachineSet\", \"VSphereDriverConfiguration\", \"VSphereMultiVCenters\", \"VSphereStaticIPs\", \"ValidatingAdmissionPolicy\"}, Disabled:[]v1.FeatureGateName{\"AWSClusterHostedDNS\", \"AdditionalRoutingCapabilities\", \"AutomatedEtcdBackup\", \"BootcNodeManagement\", \"CSIDriverSharedResource\", \"ClusterAPIInstall\", \"ClusterAPIInstallIBMCloud\", \"ClusterMonitoringConfig\", \"ConsolePluginContentSecurityPolicy\", \"DNSNameResolver\", \"DynamicResourceAllocation\", \"EtcdBackendQuota\", \"EventedPLEG\", \"Example\", \"ExternalOIDC\", \"ExternalOIDCWithUIDAndExtraClaimMappings\", \"GCPClusterHostedDNS\", \"GatewayAPI\", \"ImageStreamImportMode\", \"IngressControllerDynamicConfigurationManager\", \"InsightsConfig\", \"InsightsConfigAPI\", \"InsightsOnDemandDataGather\", \"InsightsRuntimeExtractor\", \"MachineAPIMigration\", \"MachineAPIOperatorDisableMachineHealthCheckController\", \"MachineAPIProviderOpenStack\", \"MachineConfigNodes\", \"MaxUnavailableStatefulSet\", \"MetricsCollectionProfiles\", \"MinimumKubeletVersion\", \"MixedCPUsAllocation\", \"MultiArchInstallAzure\", \"NodeSwap\", \"NutanixMultiSubnets\", \"OVNObservability\", \"OpenShiftPodSecurityAdmission\", \"PinnedImages\", \"PlatformOperators\", \"ProcMountType\", \"RouteAdvertisements\", \"RouteExternalCertificate\", \"ServiceAccountTokenNodeBinding\", \"SignatureStores\", \"SigstoreImageVerification\", \"StreamingCollectionEncodingToJSON\", \"StreamingCollectionEncodingToProtobuf\", \"TranslateStreamCloseWebsocketRequests\", \"UpgradeStatus\", \"UserNamespacesPodSecurityStandards\", \"UserNamespacesSupport\", \"VSphereMultiNetworks\", \"VolumeAttributesClass\", \"VolumeGroupSnapshot\"}}","type":"Normal"},{"namespace":"openshift-kube-controller-manager-operator","lastTimestamp":"2025-12-04T11:50:12Z","reason":"StartingNewRevision","message":"new revision 3 triggered by \"required configmap/kube-controller-manager-pod has changed,required configmap/serviceaccount-ca has changed\"","type":"Normal"},{"namespace":"openshift-kube-controller-manager-operator","lastTimestamp":"2025-12-04T11:50:13Z","reason":"OperatorVersionChanged","message":"clusteroperator/kube-controller-manager version \"kube-controller-manager\" changed from \"\" to \"1.31.13\"","type":"Normal"},{"namespace":"openshift-kube-controller-manager-operator","lastTimestamp":"2025-12-04T11:50:13Z","reason":"OperatorVersionChanged","message":"clusteroperator/kube-controller-manager version \"operator\" changed from \"\" to \"4.18.29\"","type":"Normal"},{"namespace":"openshift-kube-controller-manager-operator","lastTimestamp":"2025-12-04T11:50:13Z","reason":"OperatorStatusChanged","message":"Status for clusteroperator/kube-controller-manager changed: status.versions changed from [{\"raw-internal\" \"4.18.29\"}] to [{\"raw-internal\" \"4.18.29\"} {\"kube-controller-manager\" \"1.31.13\"} {\"operator\" \"4.18.29\"}]","type":"Normal"},{"namespace":"openshift-kube-controller-manager-operator","lastTimestamp":"2025-12-04T11:50:13Z","reason":"OperatorStatusChanged","message":"Status for clusteroperator/kube-controller-manager changed: Degraded message changed from \"NodeControllerDegraded: All master nodes are ready\" to \"StaticPodsDegraded: pod/kube-controller-manager-master-0 container \\\"cluster-policy-controller\\\" is waiting: CrashLoopBackOff: back-off 1m20s restarting failed container=cluster-policy-controller pod=kube-controller-manager-master-0_openshift-kube-controller-manager(e6b437c60bb18680f4492b00b294e872)\\nNodeControllerDegraded: All master nodes are ready\"","type":"Normal"},{"namespace":"openshift-kube-controller-manager-operator","lastTimestamp":"2025-12-04T11:50:15Z","reason":"ConfigMapCreated","message":"Created ConfigMap/recycler-config-3 -n openshift-kube-controller-manager because it was missing","type":"Normal"},{"namespace":"openshift-kube-controller-manager-operator","lastTimestamp":"2025-12-04T11:50:16Z","reason":"SecretCreated","message":"Created Secret/service-account-private-key-3 -n openshift-kube-controller-manager because it was missing","type":"Normal"},{"namespace":"openshift-kube-controller-manager-operator","lastTimestamp":"2025-12-04T11:50:17Z","reason":"SecretCreated","message":"Created Secret/serving-cert-3 -n openshift-kube-controller-manager because it was missing","type":"Normal"},{"namespace":"openshift-kube-controller-manager-operator","lastTimestamp":"2025-12-04T11:50:18Z","reason":"SecretCreated","message":"Created Secret/localhost-recovery-client-token-3 -n openshift-kube-controller-manager because it was missing","type":"Normal"},{"namespace":"openshift-kube-controller-manager-operator","lastTimestamp":"2025-12-04T11:50:19Z","reason":"RevisionTriggered","message":"new revision 3 triggered by \"required configmap/kube-controller-manager-pod has changed,required configmap/serviceaccount-ca has changed\"","type":"Normal"},{"namespace":"openshift-kube-controller-manager-operator","lastTimestamp":"2025-12-04T11:50:22Z","reason":"OperatorStatusChanged","message":"Status for clusteroperator/kube-controller-manager changed: Progressing message changed from \"NodeInstallerProgressing: 1 node is at revision 0; 0 nodes have achieved new revision 2\" to \"NodeInstallerProgressing: 1 node is at revision 0; 0 nodes have achieved new revision 3\",Available message changed from \"StaticPodsAvailable: 0 nodes are active; 1 node is at revision 0; 0 nodes have achieved new revision 2\" to \"StaticPodsAvailable: 0 nodes are active; 1 node is at revision 0; 0 nodes have achieved new revision 3\"","type":"Normal"},{"namespace":"openshift-kube-controller-manager-operator","lastTimestamp":"2025-12-04T11:50:24Z","reason":"PodCreated","message":"Created Pod/installer-3-master-0 -n openshift-kube-controller-manager because it was missing","type":"Normal"},{"namespace":"openshift-kube-controller-manager-operator","lastTimestamp":"2025-12-04T11:50:49Z","reason":"SATokenSignerControllerStuck","message":"unexpected addresses: 192.168.32.10","type":"Normal"},{"namespace":"openshift-kube-controller-manager-operator","lastTimestamp":"2025-12-04T11:55:45Z","reason":"Pulled","message":"Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a824e468cf8dd61d347e35b2ee5bc2f815666957647098e21a1bb56ff613e5b9\" already present on machine","type":"Normal"},{"namespace":"openshift-kube-controller-manager-operator","lastTimestamp":"2025-12-04T11:55:45Z","reason":"Created","message":"Created container: kube-controller-manager-operator","type":"Normal"},{"namespace":"openshift-kube-controller-manager-operator","lastTimestamp":"2025-12-04T11:55:45Z","reason":"Started","message":"Started container kube-controller-manager-operator","type":"Normal"}]}events/openshift-kube-controller-manager.json0000640000000000000000000004304115114274105020155 0ustar0000000000000000{"items":[{"namespace":"openshift-kube-controller-manager","lastTimestamp":"2025-12-04T11:38:32Z","reason":"AddedInterface","message":"Add eth0 [10.128.0.46/23] from ovn-kubernetes","type":"Normal"},{"namespace":"openshift-kube-controller-manager","lastTimestamp":"2025-12-04T11:38:33Z","reason":"Pulled","message":"Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a824e468cf8dd61d347e35b2ee5bc2f815666957647098e21a1bb56ff613e5b9\" already present on machine","type":"Normal"},{"namespace":"openshift-kube-controller-manager","lastTimestamp":"2025-12-04T11:38:33Z","reason":"Created","message":"Created container: installer","type":"Normal"},{"namespace":"openshift-kube-controller-manager","lastTimestamp":"2025-12-04T11:38:33Z","reason":"Started","message":"Started container installer","type":"Normal"},{"namespace":"openshift-kube-controller-manager","lastTimestamp":"2025-12-04T11:38:43Z","reason":"Killing","message":"Stopping container installer","type":"Normal"},{"namespace":"openshift-kube-controller-manager","lastTimestamp":"2025-12-04T11:40:22Z","reason":"FailedCreatePodSandBox","message":"Failed to create pod sandbox: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_installer-2-master-0_openshift-kube-controller-manager_38e7be62-e4f5-42ba-89f0-83aca874a092_0(07b26902c86c19476e0b1231675a3a2190e6256c4ece69045cbb7f750231e1e8): error adding pod openshift-kube-controller-manager_installer-2-master-0 to CNI network \"multus-cni-network\": plugin type=\"multus-shim\" name=\"multus-cni-network\" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:\"07b26902c86c19476e0b1231675a3a2190e6256c4ece69045cbb7f750231e1e8\" Netns:\"/var/run/netns/f85fdd93-a4af-4e9a-9d1a-34ea0b226cb0\" IfName:\"eth0\" Args:\"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-kube-controller-manager;K8S_POD_NAME=installer-2-master-0;K8S_POD_INFRA_CONTAINER_ID=07b26902c86c19476e0b1231675a3a2190e6256c4ece69045cbb7f750231e1e8;K8S_POD_UID=38e7be62-e4f5-42ba-89f0-83aca874a092\" Path:\"\" ERRORED: error configuring pod [openshift-kube-controller-manager/installer-2-master-0] networking: Multus: [openshift-kube-controller-manager/installer-2-master-0/38e7be62-e4f5-42ba-89f0-83aca874a092]: error setting the networks status: SetPodNetworkStatusAnnotation: failed to update the pod installer-2-master-0 in out of cluster comm: SetNetworkStatus: failed to update the pod installer-2-master-0 in out of cluster comm: status update failed for pod /: Get \"https://api-int.sno.openstack.lab:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/installer-2-master-0?timeout=1m0s\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)\n': StdinData: {\"binDir\":\"/var/lib/cni/bin\",\"clusterNetwork\":\"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf\",\"cniVersion\":\"0.3.1\",\"daemonSocketDir\":\"/run/multus/socket\",\"globalNamespaces\":\"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv\",\"logLevel\":\"verbose\",\"logToStderr\":true,\"name\":\"multus-cni-network\",\"namespaceIsolation\":true,\"type\":\"multus-shim\"}","type":"Warning"},{"namespace":"openshift-kube-controller-manager","lastTimestamp":"2025-12-04T11:41:23Z","reason":"FailedCreatePodSandBox","message":"Failed to create pod sandbox: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_installer-2-master-0_openshift-kube-controller-manager_38e7be62-e4f5-42ba-89f0-83aca874a092_0(acf5a98f600bd6c83a28743f2310f72c5e66e0b63da427c52d49a771c7352c34): error adding pod openshift-kube-controller-manager_installer-2-master-0 to CNI network \"multus-cni-network\": plugin type=\"multus-shim\" name=\"multus-cni-network\" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:\"acf5a98f600bd6c83a28743f2310f72c5e66e0b63da427c52d49a771c7352c34\" Netns:\"/var/run/netns/05576a5b-072e-4965-beb5-cc6ad5a5f495\" IfName:\"eth0\" Args:\"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-kube-controller-manager;K8S_POD_NAME=installer-2-master-0;K8S_POD_INFRA_CONTAINER_ID=acf5a98f600bd6c83a28743f2310f72c5e66e0b63da427c52d49a771c7352c34;K8S_POD_UID=38e7be62-e4f5-42ba-89f0-83aca874a092\" Path:\"\" ERRORED: error configuring pod [openshift-kube-controller-manager/installer-2-master-0] networking: Multus: [openshift-kube-controller-manager/installer-2-master-0/38e7be62-e4f5-42ba-89f0-83aca874a092]: error setting the networks status: SetPodNetworkStatusAnnotation: failed to update the pod installer-2-master-0 in out of cluster comm: SetNetworkStatus: failed to update the pod installer-2-master-0 in out of cluster comm: status update failed for pod /: Get \"https://api-int.sno.openstack.lab:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/installer-2-master-0?timeout=1m0s\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)\n': StdinData: {\"binDir\":\"/var/lib/cni/bin\",\"clusterNetwork\":\"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf\",\"cniVersion\":\"0.3.1\",\"daemonSocketDir\":\"/run/multus/socket\",\"globalNamespaces\":\"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv\",\"logLevel\":\"verbose\",\"logToStderr\":true,\"name\":\"multus-cni-network\",\"namespaceIsolation\":true,\"type\":\"multus-shim\"}","type":"Warning"},{"namespace":"openshift-kube-controller-manager","lastTimestamp":"2025-12-04T11:42:38Z","reason":"FailedCreatePodSandBox","message":"Failed to create pod sandbox: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_installer-2-master-0_openshift-kube-controller-manager_38e7be62-e4f5-42ba-89f0-83aca874a092_0(d1537d2edd419ed8b01c58612db0b8ec6b2e006099ec1b3cdcd856307a088d63): error adding pod openshift-kube-controller-manager_installer-2-master-0 to CNI network \"multus-cni-network\": plugin type=\"multus-shim\" name=\"multus-cni-network\" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:\"d1537d2edd419ed8b01c58612db0b8ec6b2e006099ec1b3cdcd856307a088d63\" Netns:\"/var/run/netns/5a243ddf-ba4d-40c2-894c-e6b946852288\" IfName:\"eth0\" Args:\"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-kube-controller-manager;K8S_POD_NAME=installer-2-master-0;K8S_POD_INFRA_CONTAINER_ID=d1537d2edd419ed8b01c58612db0b8ec6b2e006099ec1b3cdcd856307a088d63;K8S_POD_UID=38e7be62-e4f5-42ba-89f0-83aca874a092\" Path:\"\" ERRORED: error configuring pod [openshift-kube-controller-manager/installer-2-master-0] networking: Multus: [openshift-kube-controller-manager/installer-2-master-0/38e7be62-e4f5-42ba-89f0-83aca874a092]: error setting the networks status: SetPodNetworkStatusAnnotation: failed to update the pod installer-2-master-0 in out of cluster comm: SetNetworkStatus: failed to update the pod installer-2-master-0 in out of cluster comm: status update failed for pod /: Get \"https://api-int.sno.openstack.lab:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/installer-2-master-0?timeout=1m0s\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)\n': StdinData: {\"binDir\":\"/var/lib/cni/bin\",\"clusterNetwork\":\"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf\",\"cniVersion\":\"0.3.1\",\"daemonSocketDir\":\"/run/multus/socket\",\"globalNamespaces\":\"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv\",\"logLevel\":\"verbose\",\"logToStderr\":true,\"name\":\"multus-cni-network\",\"namespaceIsolation\":true,\"type\":\"multus-shim\"}","type":"Warning"},{"namespace":"openshift-kube-controller-manager","lastTimestamp":"2025-12-04T11:44:20Z","reason":"FailedCreatePodSandBox","message":"Failed to create pod sandbox: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_installer-2-master-0_openshift-kube-controller-manager_38e7be62-e4f5-42ba-89f0-83aca874a092_0(04ebacdbcb644ae006a00ae82dfcd15d62a4254c398f5c5c941eb7cab1b204e1): error adding pod openshift-kube-controller-manager_installer-2-master-0 to CNI network \"multus-cni-network\": plugin type=\"multus-shim\" name=\"multus-cni-network\" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:\"04ebacdbcb644ae006a00ae82dfcd15d62a4254c398f5c5c941eb7cab1b204e1\" Netns:\"/var/run/netns/7a5fa44b-229f-4be7-a3ea-eefe545a0517\" IfName:\"eth0\" Args:\"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-kube-controller-manager;K8S_POD_NAME=installer-2-master-0;K8S_POD_INFRA_CONTAINER_ID=04ebacdbcb644ae006a00ae82dfcd15d62a4254c398f5c5c941eb7cab1b204e1;K8S_POD_UID=38e7be62-e4f5-42ba-89f0-83aca874a092\" Path:\"\" ERRORED: error configuring pod [openshift-kube-controller-manager/installer-2-master-0] networking: Multus: [openshift-kube-controller-manager/installer-2-master-0/38e7be62-e4f5-42ba-89f0-83aca874a092]: error setting the networks status: SetPodNetworkStatusAnnotation: failed to update the pod installer-2-master-0 in out of cluster comm: SetNetworkStatus: failed to update the pod installer-2-master-0 in out of cluster comm: status update failed for pod /: Get \"https://api-int.sno.openstack.lab:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/installer-2-master-0?timeout=1m0s\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)\n': StdinData: {\"binDir\":\"/var/lib/cni/bin\",\"clusterNetwork\":\"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf\",\"cniVersion\":\"0.3.1\",\"daemonSocketDir\":\"/run/multus/socket\",\"globalNamespaces\":\"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv\",\"logLevel\":\"verbose\",\"logToStderr\":true,\"name\":\"multus-cni-network\",\"namespaceIsolation\":true,\"type\":\"multus-shim\"}","type":"Warning"},{"namespace":"openshift-kube-controller-manager","lastTimestamp":"2025-12-04T11:45:02Z","reason":"AddedInterface","message":"Add eth0 [10.128.0.68/23] from ovn-kubernetes","type":"Normal"},{"namespace":"openshift-kube-controller-manager","lastTimestamp":"2025-12-04T11:45:10Z","reason":"Pulled","message":"Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a824e468cf8dd61d347e35b2ee5bc2f815666957647098e21a1bb56ff613e5b9\" already present on machine","type":"Normal"},{"namespace":"openshift-kube-controller-manager","lastTimestamp":"2025-12-04T11:45:11Z","reason":"Created","message":"Created container: installer","type":"Normal"},{"namespace":"openshift-kube-controller-manager","lastTimestamp":"2025-12-04T11:45:11Z","reason":"Started","message":"Started container installer","type":"Normal"},{"namespace":"openshift-kube-controller-manager","lastTimestamp":"2025-12-04T11:45:44Z","reason":"StaticPodInstallerCompleted","message":"Successfully installed revision 2","type":"Normal"},{"namespace":"openshift-kube-controller-manager","lastTimestamp":"2025-12-04T11:45:44Z","reason":"Pulled","message":"Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:6d5891cdd7dcf7c9081de8b364b4c96446b7f946f7880fbae291a4592a198264\" already present on machine","type":"Normal"},{"namespace":"openshift-kube-controller-manager","lastTimestamp":"2025-12-04T11:45:44Z","reason":"Created","message":"Created container: kube-controller-manager","type":"Normal"},{"namespace":"openshift-kube-controller-manager","lastTimestamp":"2025-12-04T11:45:44Z","reason":"Started","message":"Started container kube-controller-manager","type":"Normal"},{"namespace":"openshift-kube-controller-manager","lastTimestamp":"2025-12-04T11:45:45Z","reason":"Started","message":"Started container kube-controller-manager-recovery-controller","type":"Normal"},{"namespace":"openshift-kube-controller-manager","lastTimestamp":"2025-12-04T11:45:45Z","reason":"Created","message":"Created container: kube-controller-manager-recovery-controller","type":"Normal"},{"namespace":"openshift-kube-controller-manager","lastTimestamp":"2025-12-04T11:45:45Z","reason":"Pulled","message":"Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a824e468cf8dd61d347e35b2ee5bc2f815666957647098e21a1bb56ff613e5b9\" already present on machine","type":"Normal"},{"namespace":"openshift-kube-controller-manager","lastTimestamp":"2025-12-04T11:45:45Z","reason":"Started","message":"Started container kube-controller-manager-cert-syncer","type":"Normal"},{"namespace":"openshift-kube-controller-manager","lastTimestamp":"2025-12-04T11:45:45Z","reason":"Created","message":"Created container: kube-controller-manager-cert-syncer","type":"Normal"},{"namespace":"openshift-kube-controller-manager","lastTimestamp":"2025-12-04T11:45:45Z","reason":"Pulled","message":"Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a824e468cf8dd61d347e35b2ee5bc2f815666957647098e21a1bb56ff613e5b9\" already present on machine","type":"Normal"},{"namespace":"openshift-kube-controller-manager","lastTimestamp":"2025-12-04T11:46:15Z","reason":"Created","message":"Created container: cluster-policy-controller","type":"Normal"},{"namespace":"openshift-kube-controller-manager","lastTimestamp":"2025-12-04T11:46:15Z","reason":"Pulled","message":"Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d64c13fe7663a0b4ae61d103b1b7598adcf317a01826f296bcb66b1a2de83c96\" already present on machine","type":"Normal"},{"namespace":"openshift-kube-controller-manager","lastTimestamp":"2025-12-04T11:46:15Z","reason":"Killing","message":"Container cluster-policy-controller failed startup probe, will be restarted","type":"Normal"},{"namespace":"openshift-kube-controller-manager","lastTimestamp":"2025-12-04T11:46:15Z","reason":"Unhealthy","message":"Startup probe failed: Get \"https://localhost:10357/healthz\": read tcp 127.0.0.1:57002-\u003e127.0.0.1:10357: read: connection reset by peer","type":"Warning"},{"namespace":"openshift-kube-controller-manager","lastTimestamp":"2025-12-04T11:46:15Z","reason":"ProbeError","message":"Startup probe error: Get \"https://localhost:10357/healthz\": read tcp 127.0.0.1:57002-\u003e127.0.0.1:10357: read: connection reset by peer\nbody: \n","type":"Warning"},{"namespace":"openshift-kube-controller-manager","lastTimestamp":"2025-12-04T11:46:15Z","reason":"Started","message":"Started container cluster-policy-controller","type":"Normal"},{"namespace":"openshift-kube-controller-manager","lastTimestamp":"2025-12-04T11:46:27Z","reason":"Unhealthy","message":"Startup probe failed: Get \"https://localhost:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)","type":"Warning"},{"namespace":"openshift-kube-controller-manager","lastTimestamp":"2025-12-04T11:46:37Z","reason":"ProbeError","message":"Startup probe error: Get \"https://localhost:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)\nbody: \n","type":"Warning"},{"namespace":"openshift-kube-controller-manager","lastTimestamp":"2025-12-04T11:50:25Z","reason":"AddedInterface","message":"Add eth0 [10.128.0.73/23] from ovn-kubernetes","type":"Normal"},{"namespace":"openshift-kube-controller-manager","lastTimestamp":"2025-12-04T11:50:25Z","reason":"Pulled","message":"Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a824e468cf8dd61d347e35b2ee5bc2f815666957647098e21a1bb56ff613e5b9\" already present on machine","type":"Normal"},{"namespace":"openshift-kube-controller-manager","lastTimestamp":"2025-12-04T11:50:25Z","reason":"Created","message":"Created container: installer","type":"Normal"},{"namespace":"openshift-kube-controller-manager","lastTimestamp":"2025-12-04T11:50:25Z","reason":"Started","message":"Started container installer","type":"Normal"},{"namespace":"openshift-kube-controller-manager","lastTimestamp":"2025-12-04T11:55:46Z","reason":"BackOff","message":"Back-off restarting failed container cluster-policy-controller in pod kube-controller-manager-master-0_openshift-kube-controller-manager(e6b437c60bb18680f4492b00b294e872)","type":"Warning"},{"namespace":"openshift-kube-controller-manager","lastTimestamp":"2025-12-04T12:00:14Z","reason":"Pulled","message":"Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d64c13fe7663a0b4ae61d103b1b7598adcf317a01826f296bcb66b1a2de83c96\" already present on machine","type":"Normal"},{"namespace":"openshift-kube-controller-manager","lastTimestamp":"2025-12-04T12:00:14Z","reason":"ProbeError","message":"Startup probe error: Get \"https://localhost:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)\nbody: \n","type":"Warning"},{"namespace":"openshift-kube-controller-manager","lastTimestamp":"2025-12-04T12:00:14Z","reason":"Unhealthy","message":"Startup probe failed: Get \"https://localhost:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)","type":"Warning"},{"namespace":"openshift-kube-controller-manager","lastTimestamp":"2025-12-04T12:00:14Z","reason":"Killing","message":"Container cluster-policy-controller failed startup probe, will be restarted","type":"Normal"},{"namespace":"openshift-kube-controller-manager","lastTimestamp":"2025-12-04T12:00:15Z","reason":"Created","message":"Created container: cluster-policy-controller","type":"Normal"},{"namespace":"openshift-kube-controller-manager","lastTimestamp":"2025-12-04T12:00:15Z","reason":"Started","message":"Started container cluster-policy-controller","type":"Normal"},{"namespace":"openshift-kube-controller-manager","lastTimestamp":"2025-12-04T12:00:24Z","reason":"ProbeError","message":"Startup probe error: Get \"https://localhost:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)\nbody: \n","type":"Warning"},{"namespace":"openshift-kube-controller-manager","lastTimestamp":"2025-12-04T12:00:24Z","reason":"Unhealthy","message":"Startup probe failed: Get \"https://localhost:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)","type":"Warning"}]}events/openshift-ingress-operator.json0000640000000000000000000005070615114274105016747 0ustar0000000000000000{"items":[{"namespace":"openshift-ingress-operator","lastTimestamp":"2025-12-04T11:31:43Z","reason":"ScalingReplicaSet","message":"Scaled up replica set ingress-operator-8649c48786 to 1","type":"Normal"},{"namespace":"openshift-ingress-operator","lastTimestamp":"2025-12-04T11:31:46Z","reason":"FailedCreate","message":"Error creating: pods \"ingress-operator-8649c48786-\" is forbidden: autoscaling.openshift.io/ManagementCPUsOverride the cluster does not have any nodes","type":"Warning"},{"namespace":"openshift-ingress-operator","lastTimestamp":"2025-12-04T11:35:47Z","reason":"FailedCreate","message":"Error creating: pods \"ingress-operator-8649c48786-\" is forbidden: autoscaling.openshift.io/ManagementCPUsOverride the cluster does not have any nodes","type":"Warning"},{"namespace":"openshift-ingress-operator","lastTimestamp":"2025-12-04T11:35:50Z","reason":"FailedScheduling","message":"0/1 nodes are available: 1 node(s) had untolerated taint {node.kubernetes.io/not-ready: }. preemption: 0/1 nodes are available: 1 Preemption is not helpful for scheduling.","type":"Warning"},{"namespace":"openshift-ingress-operator","lastTimestamp":"2025-12-04T11:35:50Z","reason":"SuccessfulCreate","message":"Created pod: ingress-operator-8649c48786-cx2b2","type":"Normal"},{"namespace":"openshift-ingress-operator","lastTimestamp":"2025-12-04T11:37:30Z","reason":"Scheduled","message":"Successfully assigned openshift-ingress-operator/ingress-operator-8649c48786-cx2b2 to master-0","type":"Normal"},{"namespace":"openshift-ingress-operator","lastTimestamp":"2025-12-04T11:37:38Z","reason":"FailedMount","message":"MountVolume.SetUp failed for volume \"metrics-tls\" : secret \"metrics-tls\" not found","type":"Warning"},{"namespace":"openshift-ingress-operator","lastTimestamp":"2025-12-04T11:37:54Z","reason":"FailedMount","message":"MountVolume.SetUp failed for volume \"metrics-tls\" : secret \"metrics-tls\" not found","type":"Warning"},{"namespace":"openshift-ingress-operator","lastTimestamp":"2025-12-04T11:38:02Z","reason":"Pulling","message":"Pulling image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:831f30660844091d6154e2674d3a9da6f34271bf8a2c40b56f7416066318742b\"","type":"Normal"},{"namespace":"openshift-ingress-operator","lastTimestamp":"2025-12-04T11:38:02Z","reason":"AddedInterface","message":"Add eth0 [10.128.0.20/23] from ovn-kubernetes","type":"Normal"},{"namespace":"openshift-ingress-operator","lastTimestamp":"2025-12-04T11:38:18Z","reason":"Pulled","message":"Successfully pulled image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:831f30660844091d6154e2674d3a9da6f34271bf8a2c40b56f7416066318742b\" in 15.761s (15.761s including waiting). Image size: 505649178 bytes.","type":"Normal"},{"namespace":"openshift-ingress-operator","lastTimestamp":"2025-12-04T11:38:19Z","reason":"Pulled","message":"Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69ffd8f8dcceedc2d6eb306cea33f8beabc1be1308cd5f4ee8b9a8e3eab9843\" already present on machine","type":"Normal"},{"namespace":"openshift-ingress-operator","lastTimestamp":"2025-12-04T11:38:19Z","reason":"Created","message":"Created container: kube-rbac-proxy","type":"Normal"},{"namespace":"openshift-ingress-operator","lastTimestamp":"2025-12-04T11:38:19Z","reason":"Started","message":"Started container kube-rbac-proxy","type":"Normal"},{"namespace":"openshift-ingress-operator","lastTimestamp":"2025-12-04T11:38:19Z","reason":"FeatureGatesInitialized","message":"FeatureGates updated to featuregates.Features{Enabled:[]v1.FeatureGateName{\"AWSEFSDriverVolumeMetrics\", \"AdminNetworkPolicy\", \"AlibabaPlatform\", \"AzureWorkloadIdentity\", \"BareMetalLoadBalancer\", \"BuildCSIVolumes\", \"ChunkSizeMiB\", \"CloudDualStackNodeIPs\", \"DisableKubeletCloudCredentialProviders\", \"GCPLabelsTags\", \"HardwareSpeed\", \"IngressControllerLBSubnetsAWS\", \"KMSv1\", \"ManagedBootImages\", \"ManagedBootImagesAWS\", \"MultiArchInstallAWS\", \"MultiArchInstallGCP\", \"NetworkDiagnosticsConfig\", \"NetworkLiveMigration\", \"NetworkSegmentation\", \"NewOLM\", \"NodeDisruptionPolicy\", \"OnClusterBuild\", \"PersistentIPsForVirtualization\", \"PrivateHostedZoneAWS\", \"SetEIPForNLBIngressController\", \"VSphereControlPlaneMachineSet\", \"VSphereDriverConfiguration\", \"VSphereMultiVCenters\", \"VSphereStaticIPs\", \"ValidatingAdmissionPolicy\"}, Disabled:[]v1.FeatureGateName{\"AWSClusterHostedDNS\", \"AdditionalRoutingCapabilities\", \"AutomatedEtcdBackup\", \"BootcNodeManagement\", \"CSIDriverSharedResource\", \"ClusterAPIInstall\", \"ClusterAPIInstallIBMCloud\", \"ClusterMonitoringConfig\", \"ConsolePluginContentSecurityPolicy\", \"DNSNameResolver\", \"DynamicResourceAllocation\", \"EtcdBackendQuota\", \"EventedPLEG\", \"Example\", \"ExternalOIDC\", \"ExternalOIDCWithUIDAndExtraClaimMappings\", \"GCPClusterHostedDNS\", \"GatewayAPI\", \"ImageStreamImportMode\", \"IngressControllerDynamicConfigurationManager\", \"InsightsConfig\", \"InsightsConfigAPI\", \"InsightsOnDemandDataGather\", \"InsightsRuntimeExtractor\", \"MachineAPIMigration\", \"MachineAPIOperatorDisableMachineHealthCheckController\", \"MachineAPIProviderOpenStack\", \"MachineConfigNodes\", \"MaxUnavailableStatefulSet\", \"MetricsCollectionProfiles\", \"MinimumKubeletVersion\", \"MixedCPUsAllocation\", \"MultiArchInstallAzure\", \"NodeSwap\", \"NutanixMultiSubnets\", \"OVNObservability\", \"OpenShiftPodSecurityAdmission\", \"PinnedImages\", \"PlatformOperators\", \"ProcMountType\", \"RouteAdvertisements\", \"RouteExternalCertificate\", \"ServiceAccountTokenNodeBinding\", \"SignatureStores\", \"SigstoreImageVerification\", \"StreamingCollectionEncodingToJSON\", \"StreamingCollectionEncodingToProtobuf\", \"TranslateStreamCloseWebsocketRequests\", \"UpgradeStatus\", \"UserNamespacesPodSecurityStandards\", \"UserNamespacesSupport\", \"VSphereMultiNetworks\", \"VolumeAttributesClass\", \"VolumeGroupSnapshot\"}}","type":"Normal"},{"namespace":"openshift-ingress-operator","lastTimestamp":"2025-12-04T11:38:20Z","reason":"CreatedWildcardCACert","message":"Created a default wildcard CA certificate","type":"Normal"},{"namespace":"openshift-ingress-operator","lastTimestamp":"2025-12-04T11:38:20Z","reason":"Admitted","message":"ingresscontroller passed validation","type":"Normal"},{"namespace":"openshift-ingress-operator","lastTimestamp":"2025-12-04T11:38:22Z","reason":"CreatedDefaultCertificate","message":"Created default wildcard certificate \"router-certs-default\"","type":"Normal"},{"namespace":"openshift-ingress-operator","lastTimestamp":"2025-12-04T11:44:54Z","reason":"FeatureGatesInitialized","message":"FeatureGates updated to featuregates.Features{Enabled:[]v1.FeatureGateName{\"AWSEFSDriverVolumeMetrics\", \"AdminNetworkPolicy\", \"AlibabaPlatform\", \"AzureWorkloadIdentity\", \"BareMetalLoadBalancer\", \"BuildCSIVolumes\", \"ChunkSizeMiB\", \"CloudDualStackNodeIPs\", \"DisableKubeletCloudCredentialProviders\", \"GCPLabelsTags\", \"HardwareSpeed\", \"IngressControllerLBSubnetsAWS\", \"KMSv1\", \"ManagedBootImages\", \"ManagedBootImagesAWS\", \"MultiArchInstallAWS\", \"MultiArchInstallGCP\", \"NetworkDiagnosticsConfig\", \"NetworkLiveMigration\", \"NetworkSegmentation\", \"NewOLM\", \"NodeDisruptionPolicy\", \"OnClusterBuild\", \"PersistentIPsForVirtualization\", \"PrivateHostedZoneAWS\", \"SetEIPForNLBIngressController\", \"VSphereControlPlaneMachineSet\", \"VSphereDriverConfiguration\", \"VSphereMultiVCenters\", \"VSphereStaticIPs\", \"ValidatingAdmissionPolicy\"}, Disabled:[]v1.FeatureGateName{\"AWSClusterHostedDNS\", \"AdditionalRoutingCapabilities\", \"AutomatedEtcdBackup\", \"BootcNodeManagement\", \"CSIDriverSharedResource\", \"ClusterAPIInstall\", \"ClusterAPIInstallIBMCloud\", \"ClusterMonitoringConfig\", \"ConsolePluginContentSecurityPolicy\", \"DNSNameResolver\", \"DynamicResourceAllocation\", \"EtcdBackendQuota\", \"EventedPLEG\", \"Example\", \"ExternalOIDC\", \"ExternalOIDCWithUIDAndExtraClaimMappings\", \"GCPClusterHostedDNS\", \"GatewayAPI\", \"ImageStreamImportMode\", \"IngressControllerDynamicConfigurationManager\", \"InsightsConfig\", \"InsightsConfigAPI\", \"InsightsOnDemandDataGather\", \"InsightsRuntimeExtractor\", \"MachineAPIMigration\", \"MachineAPIOperatorDisableMachineHealthCheckController\", \"MachineAPIProviderOpenStack\", \"MachineConfigNodes\", \"MaxUnavailableStatefulSet\", \"MetricsCollectionProfiles\", \"MinimumKubeletVersion\", \"MixedCPUsAllocation\", \"MultiArchInstallAzure\", \"NodeSwap\", \"NutanixMultiSubnets\", \"OVNObservability\", \"OpenShiftPodSecurityAdmission\", \"PinnedImages\", \"PlatformOperators\", \"ProcMountType\", \"RouteAdvertisements\", \"RouteExternalCertificate\", \"ServiceAccountTokenNodeBinding\", \"SignatureStores\", \"SigstoreImageVerification\", \"StreamingCollectionEncodingToJSON\", \"StreamingCollectionEncodingToProtobuf\", \"TranslateStreamCloseWebsocketRequests\", \"UpgradeStatus\", \"UserNamespacesPodSecurityStandards\", \"UserNamespacesSupport\", \"VSphereMultiNetworks\", \"VolumeAttributesClass\", \"VolumeGroupSnapshot\"}}","type":"Normal"},{"namespace":"openshift-ingress-operator","lastTimestamp":"2025-12-04T11:47:52Z","reason":"Pulled","message":"Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:831f30660844091d6154e2674d3a9da6f34271bf8a2c40b56f7416066318742b\" already present on machine","type":"Normal"},{"namespace":"openshift-ingress-operator","lastTimestamp":"2025-12-04T11:47:52Z","reason":"Created","message":"Created container: ingress-operator","type":"Normal"},{"namespace":"openshift-ingress-operator","lastTimestamp":"2025-12-04T11:47:52Z","reason":"Started","message":"Started container ingress-operator","type":"Normal"},{"namespace":"openshift-ingress-operator","lastTimestamp":"2025-12-04T11:47:53Z","reason":"FeatureGatesInitialized","message":"FeatureGates updated to featuregates.Features{Enabled:[]v1.FeatureGateName{\"AWSEFSDriverVolumeMetrics\", \"AdminNetworkPolicy\", \"AlibabaPlatform\", \"AzureWorkloadIdentity\", \"BareMetalLoadBalancer\", \"BuildCSIVolumes\", \"ChunkSizeMiB\", \"CloudDualStackNodeIPs\", \"DisableKubeletCloudCredentialProviders\", \"GCPLabelsTags\", \"HardwareSpeed\", \"IngressControllerLBSubnetsAWS\", \"KMSv1\", \"ManagedBootImages\", \"ManagedBootImagesAWS\", \"MultiArchInstallAWS\", \"MultiArchInstallGCP\", \"NetworkDiagnosticsConfig\", \"NetworkLiveMigration\", \"NetworkSegmentation\", \"NewOLM\", \"NodeDisruptionPolicy\", \"OnClusterBuild\", \"PersistentIPsForVirtualization\", \"PrivateHostedZoneAWS\", \"SetEIPForNLBIngressController\", \"VSphereControlPlaneMachineSet\", \"VSphereDriverConfiguration\", \"VSphereMultiVCenters\", \"VSphereStaticIPs\", \"ValidatingAdmissionPolicy\"}, Disabled:[]v1.FeatureGateName{\"AWSClusterHostedDNS\", \"AdditionalRoutingCapabilities\", \"AutomatedEtcdBackup\", \"BootcNodeManagement\", \"CSIDriverSharedResource\", \"ClusterAPIInstall\", \"ClusterAPIInstallIBMCloud\", \"ClusterMonitoringConfig\", \"ConsolePluginContentSecurityPolicy\", \"DNSNameResolver\", \"DynamicResourceAllocation\", \"EtcdBackendQuota\", \"EventedPLEG\", \"Example\", \"ExternalOIDC\", \"ExternalOIDCWithUIDAndExtraClaimMappings\", \"GCPClusterHostedDNS\", \"GatewayAPI\", \"ImageStreamImportMode\", \"IngressControllerDynamicConfigurationManager\", \"InsightsConfig\", \"InsightsConfigAPI\", \"InsightsOnDemandDataGather\", \"InsightsRuntimeExtractor\", \"MachineAPIMigration\", \"MachineAPIOperatorDisableMachineHealthCheckController\", \"MachineAPIProviderOpenStack\", \"MachineConfigNodes\", \"MaxUnavailableStatefulSet\", \"MetricsCollectionProfiles\", \"MinimumKubeletVersion\", \"MixedCPUsAllocation\", \"MultiArchInstallAzure\", \"NodeSwap\", \"NutanixMultiSubnets\", \"OVNObservability\", \"OpenShiftPodSecurityAdmission\", \"PinnedImages\", \"PlatformOperators\", \"ProcMountType\", \"RouteAdvertisements\", \"RouteExternalCertificate\", \"ServiceAccountTokenNodeBinding\", \"SignatureStores\", \"SigstoreImageVerification\", \"StreamingCollectionEncodingToJSON\", \"StreamingCollectionEncodingToProtobuf\", \"TranslateStreamCloseWebsocketRequests\", \"UpgradeStatus\", \"UserNamespacesPodSecurityStandards\", \"UserNamespacesSupport\", \"VSphereMultiNetworks\", \"VolumeAttributesClass\", \"VolumeGroupSnapshot\"}}","type":"Normal"},{"namespace":"openshift-ingress-operator","lastTimestamp":"2025-12-04T11:50:43Z","reason":"FeatureGatesInitialized","message":"FeatureGates updated to featuregates.Features{Enabled:[]v1.FeatureGateName{\"AWSEFSDriverVolumeMetrics\", \"AdminNetworkPolicy\", \"AlibabaPlatform\", \"AzureWorkloadIdentity\", \"BareMetalLoadBalancer\", \"BuildCSIVolumes\", \"ChunkSizeMiB\", \"CloudDualStackNodeIPs\", \"DisableKubeletCloudCredentialProviders\", \"GCPLabelsTags\", \"HardwareSpeed\", \"IngressControllerLBSubnetsAWS\", \"KMSv1\", \"ManagedBootImages\", \"ManagedBootImagesAWS\", \"MultiArchInstallAWS\", \"MultiArchInstallGCP\", \"NetworkDiagnosticsConfig\", \"NetworkLiveMigration\", \"NetworkSegmentation\", \"NewOLM\", \"NodeDisruptionPolicy\", \"OnClusterBuild\", \"PersistentIPsForVirtualization\", \"PrivateHostedZoneAWS\", \"SetEIPForNLBIngressController\", \"VSphereControlPlaneMachineSet\", \"VSphereDriverConfiguration\", \"VSphereMultiVCenters\", \"VSphereStaticIPs\", \"ValidatingAdmissionPolicy\"}, Disabled:[]v1.FeatureGateName{\"AWSClusterHostedDNS\", \"AdditionalRoutingCapabilities\", \"AutomatedEtcdBackup\", \"BootcNodeManagement\", \"CSIDriverSharedResource\", \"ClusterAPIInstall\", \"ClusterAPIInstallIBMCloud\", \"ClusterMonitoringConfig\", \"ConsolePluginContentSecurityPolicy\", \"DNSNameResolver\", \"DynamicResourceAllocation\", \"EtcdBackendQuota\", \"EventedPLEG\", \"Example\", \"ExternalOIDC\", \"ExternalOIDCWithUIDAndExtraClaimMappings\", \"GCPClusterHostedDNS\", \"GatewayAPI\", \"ImageStreamImportMode\", \"IngressControllerDynamicConfigurationManager\", \"InsightsConfig\", \"InsightsConfigAPI\", \"InsightsOnDemandDataGather\", \"InsightsRuntimeExtractor\", \"MachineAPIMigration\", \"MachineAPIOperatorDisableMachineHealthCheckController\", \"MachineAPIProviderOpenStack\", \"MachineConfigNodes\", \"MaxUnavailableStatefulSet\", \"MetricsCollectionProfiles\", \"MinimumKubeletVersion\", \"MixedCPUsAllocation\", \"MultiArchInstallAzure\", \"NodeSwap\", \"NutanixMultiSubnets\", \"OVNObservability\", \"OpenShiftPodSecurityAdmission\", \"PinnedImages\", \"PlatformOperators\", \"ProcMountType\", \"RouteAdvertisements\", \"RouteExternalCertificate\", \"ServiceAccountTokenNodeBinding\", \"SignatureStores\", \"SigstoreImageVerification\", \"StreamingCollectionEncodingToJSON\", \"StreamingCollectionEncodingToProtobuf\", \"TranslateStreamCloseWebsocketRequests\", \"UpgradeStatus\", \"UserNamespacesPodSecurityStandards\", \"UserNamespacesSupport\", \"VSphereMultiNetworks\", \"VolumeAttributesClass\", \"VolumeGroupSnapshot\"}}","type":"Normal"},{"namespace":"openshift-ingress-operator","lastTimestamp":"2025-12-04T11:55:39Z","reason":"FeatureGatesInitialized","message":"FeatureGates updated to featuregates.Features{Enabled:[]v1.FeatureGateName{\"AWSEFSDriverVolumeMetrics\", \"AdminNetworkPolicy\", \"AlibabaPlatform\", \"AzureWorkloadIdentity\", \"BareMetalLoadBalancer\", \"BuildCSIVolumes\", \"ChunkSizeMiB\", \"CloudDualStackNodeIPs\", \"DisableKubeletCloudCredentialProviders\", \"GCPLabelsTags\", \"HardwareSpeed\", \"IngressControllerLBSubnetsAWS\", \"KMSv1\", \"ManagedBootImages\", \"ManagedBootImagesAWS\", \"MultiArchInstallAWS\", \"MultiArchInstallGCP\", \"NetworkDiagnosticsConfig\", \"NetworkLiveMigration\", \"NetworkSegmentation\", \"NewOLM\", \"NodeDisruptionPolicy\", \"OnClusterBuild\", \"PersistentIPsForVirtualization\", \"PrivateHostedZoneAWS\", \"SetEIPForNLBIngressController\", \"VSphereControlPlaneMachineSet\", \"VSphereDriverConfiguration\", \"VSphereMultiVCenters\", \"VSphereStaticIPs\", \"ValidatingAdmissionPolicy\"}, Disabled:[]v1.FeatureGateName{\"AWSClusterHostedDNS\", \"AdditionalRoutingCapabilities\", \"AutomatedEtcdBackup\", \"BootcNodeManagement\", \"CSIDriverSharedResource\", \"ClusterAPIInstall\", \"ClusterAPIInstallIBMCloud\", \"ClusterMonitoringConfig\", \"ConsolePluginContentSecurityPolicy\", \"DNSNameResolver\", \"DynamicResourceAllocation\", \"EtcdBackendQuota\", \"EventedPLEG\", \"Example\", \"ExternalOIDC\", \"ExternalOIDCWithUIDAndExtraClaimMappings\", \"GCPClusterHostedDNS\", \"GatewayAPI\", \"ImageStreamImportMode\", \"IngressControllerDynamicConfigurationManager\", \"InsightsConfig\", \"InsightsConfigAPI\", \"InsightsOnDemandDataGather\", \"InsightsRuntimeExtractor\", \"MachineAPIMigration\", \"MachineAPIOperatorDisableMachineHealthCheckController\", \"MachineAPIProviderOpenStack\", \"MachineConfigNodes\", \"MaxUnavailableStatefulSet\", \"MetricsCollectionProfiles\", \"MinimumKubeletVersion\", \"MixedCPUsAllocation\", \"MultiArchInstallAzure\", \"NodeSwap\", \"NutanixMultiSubnets\", \"OVNObservability\", \"OpenShiftPodSecurityAdmission\", \"PinnedImages\", \"PlatformOperators\", \"ProcMountType\", \"RouteAdvertisements\", \"RouteExternalCertificate\", \"ServiceAccountTokenNodeBinding\", \"SignatureStores\", \"SigstoreImageVerification\", \"StreamingCollectionEncodingToJSON\", \"StreamingCollectionEncodingToProtobuf\", \"TranslateStreamCloseWebsocketRequests\", \"UpgradeStatus\", \"UserNamespacesPodSecurityStandards\", \"UserNamespacesSupport\", \"VSphereMultiNetworks\", \"VolumeAttributesClass\", \"VolumeGroupSnapshot\"}}","type":"Normal"},{"namespace":"openshift-ingress-operator","lastTimestamp":"2025-12-04T11:57:50Z","reason":"BackOff","message":"Back-off restarting failed container ingress-operator in pod ingress-operator-8649c48786-cx2b2_openshift-ingress-operator(b011b1f1-3235-4e20-825b-ce711c052407)","type":"Warning"},{"namespace":"openshift-ingress-operator","lastTimestamp":"2025-12-04T11:59:30Z","reason":"Started","message":"Started container ingress-operator","type":"Normal"},{"namespace":"openshift-ingress-operator","lastTimestamp":"2025-12-04T11:59:30Z","reason":"Created","message":"Created container: ingress-operator","type":"Normal"},{"namespace":"openshift-ingress-operator","lastTimestamp":"2025-12-04T11:59:30Z","reason":"Pulled","message":"Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:831f30660844091d6154e2674d3a9da6f34271bf8a2c40b56f7416066318742b\" already present on machine","type":"Normal"},{"namespace":"openshift-ingress-operator","lastTimestamp":"2025-12-04T11:59:30Z","reason":"FeatureGatesInitialized","message":"FeatureGates updated to featuregates.Features{Enabled:[]v1.FeatureGateName{\"AWSEFSDriverVolumeMetrics\", \"AdminNetworkPolicy\", \"AlibabaPlatform\", \"AzureWorkloadIdentity\", \"BareMetalLoadBalancer\", \"BuildCSIVolumes\", \"ChunkSizeMiB\", \"CloudDualStackNodeIPs\", \"DisableKubeletCloudCredentialProviders\", \"GCPLabelsTags\", \"HardwareSpeed\", \"IngressControllerLBSubnetsAWS\", \"KMSv1\", \"ManagedBootImages\", \"ManagedBootImagesAWS\", \"MultiArchInstallAWS\", \"MultiArchInstallGCP\", \"NetworkDiagnosticsConfig\", \"NetworkLiveMigration\", \"NetworkSegmentation\", \"NewOLM\", \"NodeDisruptionPolicy\", \"OnClusterBuild\", \"PersistentIPsForVirtualization\", \"PrivateHostedZoneAWS\", \"SetEIPForNLBIngressController\", \"VSphereControlPlaneMachineSet\", \"VSphereDriverConfiguration\", \"VSphereMultiVCenters\", \"VSphereStaticIPs\", \"ValidatingAdmissionPolicy\"}, Disabled:[]v1.FeatureGateName{\"AWSClusterHostedDNS\", \"AdditionalRoutingCapabilities\", \"AutomatedEtcdBackup\", \"BootcNodeManagement\", \"CSIDriverSharedResource\", \"ClusterAPIInstall\", \"ClusterAPIInstallIBMCloud\", \"ClusterMonitoringConfig\", \"ConsolePluginContentSecurityPolicy\", \"DNSNameResolver\", \"DynamicResourceAllocation\", \"EtcdBackendQuota\", \"EventedPLEG\", \"Example\", \"ExternalOIDC\", \"ExternalOIDCWithUIDAndExtraClaimMappings\", \"GCPClusterHostedDNS\", \"GatewayAPI\", \"ImageStreamImportMode\", \"IngressControllerDynamicConfigurationManager\", \"InsightsConfig\", \"InsightsConfigAPI\", \"InsightsOnDemandDataGather\", \"InsightsRuntimeExtractor\", \"MachineAPIMigration\", \"MachineAPIOperatorDisableMachineHealthCheckController\", \"MachineAPIProviderOpenStack\", \"MachineConfigNodes\", \"MaxUnavailableStatefulSet\", \"MetricsCollectionProfiles\", \"MinimumKubeletVersion\", \"MixedCPUsAllocation\", \"MultiArchInstallAzure\", \"NodeSwap\", \"NutanixMultiSubnets\", \"OVNObservability\", \"OpenShiftPodSecurityAdmission\", \"PinnedImages\", \"PlatformOperators\", \"ProcMountType\", \"RouteAdvertisements\", \"RouteExternalCertificate\", \"ServiceAccountTokenNodeBinding\", \"SignatureStores\", \"SigstoreImageVerification\", \"StreamingCollectionEncodingToJSON\", \"StreamingCollectionEncodingToProtobuf\", \"TranslateStreamCloseWebsocketRequests\", \"UpgradeStatus\", \"UserNamespacesPodSecurityStandards\", \"UserNamespacesSupport\", \"VSphereMultiNetworks\", \"VolumeAttributesClass\", \"VolumeGroupSnapshot\"}}","type":"Normal"}]}events/openshift-image-registry.json0000640000000000000000000002755715114274105016404 0ustar0000000000000000{"items":[{"namespace":"openshift-image-registry","lastTimestamp":"2025-12-04T11:31:45Z","reason":"ScalingReplicaSet","message":"Scaled up replica set cluster-image-registry-operator-6fb9f88b7 to 1","type":"Normal"},{"namespace":"openshift-image-registry","lastTimestamp":"2025-12-04T11:31:46Z","reason":"FailedCreate","message":"Error creating: pods \"cluster-image-registry-operator-6fb9f88b7-\" is forbidden: autoscaling.openshift.io/ManagementCPUsOverride the cluster does not have any nodes","type":"Warning"},{"namespace":"openshift-image-registry","lastTimestamp":"2025-12-04T11:35:47Z","reason":"FailedCreate","message":"Error creating: pods \"cluster-image-registry-operator-6fb9f88b7-\" is forbidden: autoscaling.openshift.io/ManagementCPUsOverride the cluster does not have any nodes","type":"Warning"},{"namespace":"openshift-image-registry","lastTimestamp":"2025-12-04T11:35:49Z","reason":"FailedScheduling","message":"0/1 nodes are available: 1 node(s) had untolerated taint {node.kubernetes.io/not-ready: }. preemption: 0/1 nodes are available: 1 Preemption is not helpful for scheduling.","type":"Warning"},{"namespace":"openshift-image-registry","lastTimestamp":"2025-12-04T11:35:49Z","reason":"SuccessfulCreate","message":"Created pod: cluster-image-registry-operator-6fb9f88b7-tgvfl","type":"Normal"},{"namespace":"openshift-image-registry","lastTimestamp":"2025-12-04T11:37:30Z","reason":"Scheduled","message":"Successfully assigned openshift-image-registry/cluster-image-registry-operator-6fb9f88b7-tgvfl to master-0","type":"Normal"},{"namespace":"openshift-image-registry","lastTimestamp":"2025-12-04T11:37:38Z","reason":"FailedMount","message":"MountVolume.SetUp failed for volume \"image-registry-operator-tls\" : secret \"image-registry-operator-tls\" not found","type":"Warning"},{"namespace":"openshift-image-registry","lastTimestamp":"2025-12-04T11:37:54Z","reason":"FailedMount","message":"MountVolume.SetUp failed for volume \"image-registry-operator-tls\" : secret \"image-registry-operator-tls\" not found","type":"Warning"},{"namespace":"openshift-image-registry","lastTimestamp":"2025-12-04T11:38:02Z","reason":"Pulling","message":"Pulling image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:aa24edce3d740f84c40018e94cdbf2bc7375268d13d57c2d664e43a46ccea3fc\"","type":"Normal"},{"namespace":"openshift-image-registry","lastTimestamp":"2025-12-04T11:38:02Z","reason":"AddedInterface","message":"Add eth0 [10.128.0.24/23] from ovn-kubernetes","type":"Normal"},{"namespace":"openshift-image-registry","lastTimestamp":"2025-12-04T11:38:18Z","reason":"Pulled","message":"Successfully pulled image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:aa24edce3d740f84c40018e94cdbf2bc7375268d13d57c2d664e43a46ccea3fc\" in 15.79s (15.79s including waiting). Image size: 543227406 bytes.","type":"Normal"},{"namespace":"openshift-image-registry","lastTimestamp":"2025-12-04T11:38:19Z","reason":"FeatureGatesInitialized","message":"FeatureGates updated to featuregates.Features{Enabled:[]v1.FeatureGateName{\"AWSEFSDriverVolumeMetrics\", \"AdminNetworkPolicy\", \"AlibabaPlatform\", \"AzureWorkloadIdentity\", \"BareMetalLoadBalancer\", \"BuildCSIVolumes\", \"ChunkSizeMiB\", \"CloudDualStackNodeIPs\", \"DisableKubeletCloudCredentialProviders\", \"GCPLabelsTags\", \"HardwareSpeed\", \"IngressControllerLBSubnetsAWS\", \"KMSv1\", \"ManagedBootImages\", \"ManagedBootImagesAWS\", \"MultiArchInstallAWS\", \"MultiArchInstallGCP\", \"NetworkDiagnosticsConfig\", \"NetworkLiveMigration\", \"NetworkSegmentation\", \"NewOLM\", \"NodeDisruptionPolicy\", \"OnClusterBuild\", \"PersistentIPsForVirtualization\", \"PrivateHostedZoneAWS\", \"SetEIPForNLBIngressController\", \"VSphereControlPlaneMachineSet\", \"VSphereDriverConfiguration\", \"VSphereMultiVCenters\", \"VSphereStaticIPs\", \"ValidatingAdmissionPolicy\"}, Disabled:[]v1.FeatureGateName{\"AWSClusterHostedDNS\", \"AdditionalRoutingCapabilities\", \"AutomatedEtcdBackup\", \"BootcNodeManagement\", \"CSIDriverSharedResource\", \"ClusterAPIInstall\", \"ClusterAPIInstallIBMCloud\", \"ClusterMonitoringConfig\", \"ConsolePluginContentSecurityPolicy\", \"DNSNameResolver\", \"DynamicResourceAllocation\", \"EtcdBackendQuota\", \"EventedPLEG\", \"Example\", \"ExternalOIDC\", \"ExternalOIDCWithUIDAndExtraClaimMappings\", \"GCPClusterHostedDNS\", \"GatewayAPI\", \"ImageStreamImportMode\", \"IngressControllerDynamicConfigurationManager\", \"InsightsConfig\", \"InsightsConfigAPI\", \"InsightsOnDemandDataGather\", \"InsightsRuntimeExtractor\", \"MachineAPIMigration\", \"MachineAPIOperatorDisableMachineHealthCheckController\", \"MachineAPIProviderOpenStack\", \"MachineConfigNodes\", \"MaxUnavailableStatefulSet\", \"MetricsCollectionProfiles\", \"MinimumKubeletVersion\", \"MixedCPUsAllocation\", \"MultiArchInstallAzure\", \"NodeSwap\", \"NutanixMultiSubnets\", \"OVNObservability\", \"OpenShiftPodSecurityAdmission\", \"PinnedImages\", \"PlatformOperators\", \"ProcMountType\", \"RouteAdvertisements\", \"RouteExternalCertificate\", \"ServiceAccountTokenNodeBinding\", \"SignatureStores\", \"SigstoreImageVerification\", \"StreamingCollectionEncodingToJSON\", \"StreamingCollectionEncodingToProtobuf\", \"TranslateStreamCloseWebsocketRequests\", \"UpgradeStatus\", \"UserNamespacesPodSecurityStandards\", \"UserNamespacesSupport\", \"VSphereMultiNetworks\", \"VolumeAttributesClass\", \"VolumeGroupSnapshot\"}}","type":"Normal"},{"namespace":"openshift-image-registry","lastTimestamp":"2025-12-04T11:38:19Z","reason":"LeaderElection","message":"cluster-image-registry-operator-6fb9f88b7-tgvfl_f86fd9fc-9a0b-42c8-9102-96ca48ea8651 became leader","type":"Normal"},{"namespace":"openshift-image-registry","lastTimestamp":"2025-12-04T11:50:34Z","reason":"FeatureGatesInitialized","message":"FeatureGates updated to featuregates.Features{Enabled:[]v1.FeatureGateName{\"AWSEFSDriverVolumeMetrics\", \"AdminNetworkPolicy\", \"AlibabaPlatform\", \"AzureWorkloadIdentity\", \"BareMetalLoadBalancer\", \"BuildCSIVolumes\", \"ChunkSizeMiB\", \"CloudDualStackNodeIPs\", \"DisableKubeletCloudCredentialProviders\", \"GCPLabelsTags\", \"HardwareSpeed\", \"IngressControllerLBSubnetsAWS\", \"KMSv1\", \"ManagedBootImages\", \"ManagedBootImagesAWS\", \"MultiArchInstallAWS\", \"MultiArchInstallGCP\", \"NetworkDiagnosticsConfig\", \"NetworkLiveMigration\", \"NetworkSegmentation\", \"NewOLM\", \"NodeDisruptionPolicy\", \"OnClusterBuild\", \"PersistentIPsForVirtualization\", \"PrivateHostedZoneAWS\", \"SetEIPForNLBIngressController\", \"VSphereControlPlaneMachineSet\", \"VSphereDriverConfiguration\", \"VSphereMultiVCenters\", \"VSphereStaticIPs\", \"ValidatingAdmissionPolicy\"}, Disabled:[]v1.FeatureGateName{\"AWSClusterHostedDNS\", \"AdditionalRoutingCapabilities\", \"AutomatedEtcdBackup\", \"BootcNodeManagement\", \"CSIDriverSharedResource\", \"ClusterAPIInstall\", \"ClusterAPIInstallIBMCloud\", \"ClusterMonitoringConfig\", \"ConsolePluginContentSecurityPolicy\", \"DNSNameResolver\", \"DynamicResourceAllocation\", \"EtcdBackendQuota\", \"EventedPLEG\", \"Example\", \"ExternalOIDC\", \"ExternalOIDCWithUIDAndExtraClaimMappings\", \"GCPClusterHostedDNS\", \"GatewayAPI\", \"ImageStreamImportMode\", \"IngressControllerDynamicConfigurationManager\", \"InsightsConfig\", \"InsightsConfigAPI\", \"InsightsOnDemandDataGather\", \"InsightsRuntimeExtractor\", \"MachineAPIMigration\", \"MachineAPIOperatorDisableMachineHealthCheckController\", \"MachineAPIProviderOpenStack\", \"MachineConfigNodes\", \"MaxUnavailableStatefulSet\", \"MetricsCollectionProfiles\", \"MinimumKubeletVersion\", \"MixedCPUsAllocation\", \"MultiArchInstallAzure\", \"NodeSwap\", \"NutanixMultiSubnets\", \"OVNObservability\", \"OpenShiftPodSecurityAdmission\", \"PinnedImages\", \"PlatformOperators\", \"ProcMountType\", \"RouteAdvertisements\", \"RouteExternalCertificate\", \"ServiceAccountTokenNodeBinding\", \"SignatureStores\", \"SigstoreImageVerification\", \"StreamingCollectionEncodingToJSON\", \"StreamingCollectionEncodingToProtobuf\", \"TranslateStreamCloseWebsocketRequests\", \"UpgradeStatus\", \"UserNamespacesPodSecurityStandards\", \"UserNamespacesSupport\", \"VSphereMultiNetworks\", \"VolumeAttributesClass\", \"VolumeGroupSnapshot\"}}","type":"Normal"},{"namespace":"openshift-image-registry","lastTimestamp":"2025-12-04T11:50:34Z","reason":"LeaderElection","message":"cluster-image-registry-operator-6fb9f88b7-tgvfl_fb72665d-b1de-4acb-9306-ef17501cb970 became leader","type":"Normal"},{"namespace":"openshift-image-registry","lastTimestamp":"2025-12-04T11:55:47Z","reason":"Pulled","message":"Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:aa24edce3d740f84c40018e94cdbf2bc7375268d13d57c2d664e43a46ccea3fc\" already present on machine","type":"Normal"},{"namespace":"openshift-image-registry","lastTimestamp":"2025-12-04T11:55:47Z","reason":"Started","message":"Started container cluster-image-registry-operator","type":"Normal"},{"namespace":"openshift-image-registry","lastTimestamp":"2025-12-04T11:55:47Z","reason":"Created","message":"Created container: cluster-image-registry-operator","type":"Normal"},{"namespace":"openshift-image-registry","lastTimestamp":"2025-12-04T11:56:11Z","reason":"FeatureGatesInitialized","message":"FeatureGates updated to featuregates.Features{Enabled:[]v1.FeatureGateName{\"AWSEFSDriverVolumeMetrics\", \"AdminNetworkPolicy\", \"AlibabaPlatform\", \"AzureWorkloadIdentity\", \"BareMetalLoadBalancer\", \"BuildCSIVolumes\", \"ChunkSizeMiB\", \"CloudDualStackNodeIPs\", \"DisableKubeletCloudCredentialProviders\", \"GCPLabelsTags\", \"HardwareSpeed\", \"IngressControllerLBSubnetsAWS\", \"KMSv1\", \"ManagedBootImages\", \"ManagedBootImagesAWS\", \"MultiArchInstallAWS\", \"MultiArchInstallGCP\", \"NetworkDiagnosticsConfig\", \"NetworkLiveMigration\", \"NetworkSegmentation\", \"NewOLM\", \"NodeDisruptionPolicy\", \"OnClusterBuild\", \"PersistentIPsForVirtualization\", \"PrivateHostedZoneAWS\", \"SetEIPForNLBIngressController\", \"VSphereControlPlaneMachineSet\", \"VSphereDriverConfiguration\", \"VSphereMultiVCenters\", \"VSphereStaticIPs\", \"ValidatingAdmissionPolicy\"}, Disabled:[]v1.FeatureGateName{\"AWSClusterHostedDNS\", \"AdditionalRoutingCapabilities\", \"AutomatedEtcdBackup\", \"BootcNodeManagement\", \"CSIDriverSharedResource\", \"ClusterAPIInstall\", \"ClusterAPIInstallIBMCloud\", \"ClusterMonitoringConfig\", \"ConsolePluginContentSecurityPolicy\", \"DNSNameResolver\", \"DynamicResourceAllocation\", \"EtcdBackendQuota\", \"EventedPLEG\", \"Example\", \"ExternalOIDC\", \"ExternalOIDCWithUIDAndExtraClaimMappings\", \"GCPClusterHostedDNS\", \"GatewayAPI\", \"ImageStreamImportMode\", \"IngressControllerDynamicConfigurationManager\", \"InsightsConfig\", \"InsightsConfigAPI\", \"InsightsOnDemandDataGather\", \"InsightsRuntimeExtractor\", \"MachineAPIMigration\", \"MachineAPIOperatorDisableMachineHealthCheckController\", \"MachineAPIProviderOpenStack\", \"MachineConfigNodes\", \"MaxUnavailableStatefulSet\", \"MetricsCollectionProfiles\", \"MinimumKubeletVersion\", \"MixedCPUsAllocation\", \"MultiArchInstallAzure\", \"NodeSwap\", \"NutanixMultiSubnets\", \"OVNObservability\", \"OpenShiftPodSecurityAdmission\", \"PinnedImages\", \"PlatformOperators\", \"ProcMountType\", \"RouteAdvertisements\", \"RouteExternalCertificate\", \"ServiceAccountTokenNodeBinding\", \"SignatureStores\", \"SigstoreImageVerification\", \"StreamingCollectionEncodingToJSON\", \"StreamingCollectionEncodingToProtobuf\", \"TranslateStreamCloseWebsocketRequests\", \"UpgradeStatus\", \"UserNamespacesPodSecurityStandards\", \"UserNamespacesSupport\", \"VSphereMultiNetworks\", \"VolumeAttributesClass\", \"VolumeGroupSnapshot\"}}","type":"Normal"},{"namespace":"openshift-image-registry","lastTimestamp":"2025-12-04T11:56:11Z","reason":"LeaderElection","message":"cluster-image-registry-operator-6fb9f88b7-tgvfl_0434264d-558a-4723-99f2-056d66af3bf9 became leader","type":"Normal"},{"namespace":"openshift-image-registry","lastTimestamp":"2025-12-04T12:01:39Z","reason":"DaemonSetCreated","message":"Created DaemonSet.apps/node-ca -n openshift-image-registry because it was missing","type":"Normal"}]}events/openshift-dns.json0000640000000000000000000000605115114274105014222 0ustar0000000000000000{"items":[{"namespace":"openshift-dns","lastTimestamp":"2025-12-04T11:38:20Z","reason":"SuccessfulCreate","message":"Created pod: dns-default-xg7vh","type":"Normal"},{"namespace":"openshift-dns","lastTimestamp":"2025-12-04T11:38:20Z","reason":"Scheduled","message":"Successfully assigned openshift-dns/dns-default-xg7vh to master-0","type":"Normal"},{"namespace":"openshift-dns","lastTimestamp":"2025-12-04T11:38:21Z","reason":"FailedMount","message":"MountVolume.SetUp failed for volume \"metrics-tls\" : secret \"dns-default-metrics-tls\" not found","type":"Warning"},{"namespace":"openshift-dns","lastTimestamp":"2025-12-04T11:38:21Z","reason":"SuccessfulCreate","message":"Created pod: node-resolver-qq64m","type":"Normal"},{"namespace":"openshift-dns","lastTimestamp":"2025-12-04T11:38:21Z","reason":"Scheduled","message":"Successfully assigned openshift-dns/node-resolver-qq64m to master-0","type":"Normal"},{"namespace":"openshift-dns","lastTimestamp":"2025-12-04T11:38:22Z","reason":"AddedInterface","message":"Add eth0 [10.128.0.43/23] from ovn-kubernetes","type":"Normal"},{"namespace":"openshift-dns","lastTimestamp":"2025-12-04T11:38:22Z","reason":"Pulling","message":"Pulling image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:eb928c13a46d3fb45f4a881892d023a92d610a5430be0ffd916aaf8da8e7d297\"","type":"Normal"},{"namespace":"openshift-dns","lastTimestamp":"2025-12-04T11:38:22Z","reason":"Pulled","message":"Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:79f99fd6cce984287932edf0d009660bb488d663081f3d62ec3b23bc8bfbf6c2\" already present on machine","type":"Normal"},{"namespace":"openshift-dns","lastTimestamp":"2025-12-04T11:38:32Z","reason":"Pulled","message":"Successfully pulled image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:eb928c13a46d3fb45f4a881892d023a92d610a5430be0ffd916aaf8da8e7d297\" in 9.574s (9.574s including waiting). Image size: 478642572 bytes.","type":"Normal"},{"namespace":"openshift-dns","lastTimestamp":"2025-12-04T11:38:32Z","reason":"Created","message":"Created container: dns-node-resolver","type":"Normal"},{"namespace":"openshift-dns","lastTimestamp":"2025-12-04T11:38:32Z","reason":"Started","message":"Started container dns-node-resolver","type":"Normal"},{"namespace":"openshift-dns","lastTimestamp":"2025-12-04T11:38:33Z","reason":"Started","message":"Started container kube-rbac-proxy","type":"Normal"},{"namespace":"openshift-dns","lastTimestamp":"2025-12-04T11:38:33Z","reason":"Created","message":"Created container: kube-rbac-proxy","type":"Normal"},{"namespace":"openshift-dns","lastTimestamp":"2025-12-04T11:38:33Z","reason":"Pulled","message":"Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69ffd8f8dcceedc2d6eb306cea33f8beabc1be1308cd5f4ee8b9a8e3eab9843\" already present on machine","type":"Normal"},{"namespace":"openshift-dns","lastTimestamp":"2025-12-04T11:38:33Z","reason":"Started","message":"Started container dns","type":"Normal"},{"namespace":"openshift-dns","lastTimestamp":"2025-12-04T11:38:33Z","reason":"Created","message":"Created container: dns","type":"Normal"}]}events/openshift-dns-operator.json0000640000000000000000000001402615114274105016054 0ustar0000000000000000{"items":[{"namespace":"openshift-dns-operator","lastTimestamp":"2025-12-04T11:31:29Z","reason":"ScalingReplicaSet","message":"Scaled up replica set dns-operator-7c56cf9b74 to 1","type":"Normal"},{"namespace":"openshift-dns-operator","lastTimestamp":"2025-12-04T11:31:40Z","reason":"FailedCreate","message":"Error creating: pods \"dns-operator-7c56cf9b74-\" is forbidden: autoscaling.openshift.io/ManagementCPUsOverride the cluster does not have any nodes","type":"Warning"},{"namespace":"openshift-dns-operator","lastTimestamp":"2025-12-04T11:35:47Z","reason":"FailedCreate","message":"Error creating: pods \"dns-operator-7c56cf9b74-\" is forbidden: autoscaling.openshift.io/ManagementCPUsOverride the cluster does not have any nodes","type":"Warning"},{"namespace":"openshift-dns-operator","lastTimestamp":"2025-12-04T11:35:48Z","reason":"FailedScheduling","message":"0/1 nodes are available: 1 node(s) had untolerated taint {node.kubernetes.io/not-ready: }. preemption: 0/1 nodes are available: 1 Preemption is not helpful for scheduling.","type":"Warning"},{"namespace":"openshift-dns-operator","lastTimestamp":"2025-12-04T11:35:48Z","reason":"SuccessfulCreate","message":"Created pod: dns-operator-7c56cf9b74-xz27r","type":"Normal"},{"namespace":"openshift-dns-operator","lastTimestamp":"2025-12-04T11:37:30Z","reason":"Scheduled","message":"Successfully assigned openshift-dns-operator/dns-operator-7c56cf9b74-xz27r to master-0","type":"Normal"},{"namespace":"openshift-dns-operator","lastTimestamp":"2025-12-04T11:37:38Z","reason":"FailedMount","message":"MountVolume.SetUp failed for volume \"metrics-tls\" : secret \"metrics-tls\" not found","type":"Warning"},{"namespace":"openshift-dns-operator","lastTimestamp":"2025-12-04T11:37:54Z","reason":"FailedMount","message":"MountVolume.SetUp failed for volume \"metrics-tls\" : secret \"metrics-tls\" not found","type":"Warning"},{"namespace":"openshift-dns-operator","lastTimestamp":"2025-12-04T11:38:02Z","reason":"AddedInterface","message":"Add eth0 [10.128.0.9/23] from ovn-kubernetes","type":"Normal"},{"namespace":"openshift-dns-operator","lastTimestamp":"2025-12-04T11:38:02Z","reason":"Pulling","message":"Pulling image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3c1edf52f70bf9b1d1457e0c4111bc79cdaa1edd659ddbdb9d8176eff8b46956\"","type":"Normal"},{"namespace":"openshift-dns-operator","lastTimestamp":"2025-12-04T11:38:18Z","reason":"Pulled","message":"Successfully pulled image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3c1edf52f70bf9b1d1457e0c4111bc79cdaa1edd659ddbdb9d8176eff8b46956\" in 15.763s (15.763s including waiting). Image size: 462727837 bytes.","type":"Normal"},{"namespace":"openshift-dns-operator","lastTimestamp":"2025-12-04T11:38:19Z","reason":"Pulled","message":"Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69ffd8f8dcceedc2d6eb306cea33f8beabc1be1308cd5f4ee8b9a8e3eab9843\" already present on machine","type":"Normal"},{"namespace":"openshift-dns-operator","lastTimestamp":"2025-12-04T11:38:19Z","reason":"Started","message":"Started container dns-operator","type":"Normal"},{"namespace":"openshift-dns-operator","lastTimestamp":"2025-12-04T11:38:19Z","reason":"Created","message":"Created container: dns-operator","type":"Normal"},{"namespace":"openshift-dns-operator","lastTimestamp":"2025-12-04T11:38:19Z","reason":"FeatureGatesInitialized","message":"FeatureGates updated to featuregates.Features{Enabled:[]v1.FeatureGateName{\"AWSEFSDriverVolumeMetrics\", \"AdminNetworkPolicy\", \"AlibabaPlatform\", \"AzureWorkloadIdentity\", \"BareMetalLoadBalancer\", \"BuildCSIVolumes\", \"ChunkSizeMiB\", \"CloudDualStackNodeIPs\", \"DisableKubeletCloudCredentialProviders\", \"GCPLabelsTags\", \"HardwareSpeed\", \"IngressControllerLBSubnetsAWS\", \"KMSv1\", \"ManagedBootImages\", \"ManagedBootImagesAWS\", \"MultiArchInstallAWS\", \"MultiArchInstallGCP\", \"NetworkDiagnosticsConfig\", \"NetworkLiveMigration\", \"NetworkSegmentation\", \"NewOLM\", \"NodeDisruptionPolicy\", \"OnClusterBuild\", \"PersistentIPsForVirtualization\", \"PrivateHostedZoneAWS\", \"SetEIPForNLBIngressController\", \"VSphereControlPlaneMachineSet\", \"VSphereDriverConfiguration\", \"VSphereMultiVCenters\", \"VSphereStaticIPs\", \"ValidatingAdmissionPolicy\"}, Disabled:[]v1.FeatureGateName{\"AWSClusterHostedDNS\", \"AdditionalRoutingCapabilities\", \"AutomatedEtcdBackup\", \"BootcNodeManagement\", \"CSIDriverSharedResource\", \"ClusterAPIInstall\", \"ClusterAPIInstallIBMCloud\", \"ClusterMonitoringConfig\", \"ConsolePluginContentSecurityPolicy\", \"DNSNameResolver\", \"DynamicResourceAllocation\", \"EtcdBackendQuota\", \"EventedPLEG\", \"Example\", \"ExternalOIDC\", \"ExternalOIDCWithUIDAndExtraClaimMappings\", \"GCPClusterHostedDNS\", \"GatewayAPI\", \"ImageStreamImportMode\", \"IngressControllerDynamicConfigurationManager\", \"InsightsConfig\", \"InsightsConfigAPI\", \"InsightsOnDemandDataGather\", \"InsightsRuntimeExtractor\", \"MachineAPIMigration\", \"MachineAPIOperatorDisableMachineHealthCheckController\", \"MachineAPIProviderOpenStack\", \"MachineConfigNodes\", \"MaxUnavailableStatefulSet\", \"MetricsCollectionProfiles\", \"MinimumKubeletVersion\", \"MixedCPUsAllocation\", \"MultiArchInstallAzure\", \"NodeSwap\", \"NutanixMultiSubnets\", \"OVNObservability\", \"OpenShiftPodSecurityAdmission\", \"PinnedImages\", \"PlatformOperators\", \"ProcMountType\", \"RouteAdvertisements\", \"RouteExternalCertificate\", \"ServiceAccountTokenNodeBinding\", \"SignatureStores\", \"SigstoreImageVerification\", \"StreamingCollectionEncodingToJSON\", \"StreamingCollectionEncodingToProtobuf\", \"TranslateStreamCloseWebsocketRequests\", \"UpgradeStatus\", \"UserNamespacesPodSecurityStandards\", \"UserNamespacesSupport\", \"VSphereMultiNetworks\", \"VolumeAttributesClass\", \"VolumeGroupSnapshot\"}}","type":"Normal"},{"namespace":"openshift-dns-operator","lastTimestamp":"2025-12-04T11:38:20Z","reason":"Created","message":"Created container: kube-rbac-proxy","type":"Normal"},{"namespace":"openshift-dns-operator","lastTimestamp":"2025-12-04T11:38:20Z","reason":"Started","message":"Started container kube-rbac-proxy","type":"Normal"}]}events/openshift-oauth-apiserver.json0000640000000000000000000000444115114274105016555 0ustar0000000000000000{"items":[{"namespace":"openshift-oauth-apiserver","lastTimestamp":"2025-12-04T11:38:18Z","reason":"Scheduled","message":"Successfully assigned openshift-oauth-apiserver/apiserver-7467446c4b-dlj7g to master-0","type":"Normal"},{"namespace":"openshift-oauth-apiserver","lastTimestamp":"2025-12-04T11:38:18Z","reason":"SuccessfulCreate","message":"Created pod: apiserver-7467446c4b-dlj7g","type":"Normal"},{"namespace":"openshift-oauth-apiserver","lastTimestamp":"2025-12-04T11:38:18Z","reason":"ScalingReplicaSet","message":"Scaled up replica set apiserver-7467446c4b to 1","type":"Normal"},{"namespace":"openshift-oauth-apiserver","lastTimestamp":"2025-12-04T11:38:19Z","reason":"AddedInterface","message":"Add eth0 [10.128.0.39/23] from ovn-kubernetes","type":"Normal"},{"namespace":"openshift-oauth-apiserver","lastTimestamp":"2025-12-04T11:38:20Z","reason":"Pulling","message":"Pulling image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:91af633e585621630c40d14f188e37d36b44678d0a59e582d850bf8d593d3a0c\"","type":"Normal"},{"namespace":"openshift-oauth-apiserver","lastTimestamp":"2025-12-04T11:38:32Z","reason":"Pulled","message":"Successfully pulled image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:91af633e585621630c40d14f188e37d36b44678d0a59e582d850bf8d593d3a0c\" in 12.203s (12.203s including waiting). Image size: 499798563 bytes.","type":"Normal"},{"namespace":"openshift-oauth-apiserver","lastTimestamp":"2025-12-04T11:38:33Z","reason":"Created","message":"Created container: fix-audit-permissions","type":"Normal"},{"namespace":"openshift-oauth-apiserver","lastTimestamp":"2025-12-04T11:38:33Z","reason":"Started","message":"Started container fix-audit-permissions","type":"Normal"},{"namespace":"openshift-oauth-apiserver","lastTimestamp":"2025-12-04T11:38:33Z","reason":"Pulled","message":"Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:91af633e585621630c40d14f188e37d36b44678d0a59e582d850bf8d593d3a0c\" already present on machine","type":"Normal"},{"namespace":"openshift-oauth-apiserver","lastTimestamp":"2025-12-04T11:38:34Z","reason":"Created","message":"Created container: oauth-apiserver","type":"Normal"},{"namespace":"openshift-oauth-apiserver","lastTimestamp":"2025-12-04T11:38:34Z","reason":"Started","message":"Started container oauth-apiserver","type":"Normal"}]}events/openshift-ingress.json0000640000000000000000000000124015114274105015103 0ustar0000000000000000{"items":[{"namespace":"openshift-ingress","lastTimestamp":"2025-12-04T11:38:21Z","reason":"FailedScheduling","message":"0/1 nodes are available: 1 node(s) had untolerated taint {node-role.kubernetes.io/master: }. preemption: 0/1 nodes are available: 1 Preemption is not helpful for scheduling.","type":"Warning"},{"namespace":"openshift-ingress","lastTimestamp":"2025-12-04T11:38:21Z","reason":"SuccessfulCreate","message":"Created pod: router-default-5465c8b4db-58d52","type":"Normal"},{"namespace":"openshift-ingress","lastTimestamp":"2025-12-04T11:38:21Z","reason":"ScalingReplicaSet","message":"Scaled up replica set router-default-5465c8b4db to 1","type":"Normal"}]}events/openshift-authentication-operator.json0000640000000000000000000026143115114274105020313 0ustar0000000000000000{"items":[{"namespace":"openshift-authentication-operator","lastTimestamp":"2025-12-04T11:31:32Z","reason":"ScalingReplicaSet","message":"Scaled up replica set authentication-operator-6c968fdfdf to 1","type":"Normal"},{"namespace":"openshift-authentication-operator","lastTimestamp":"2025-12-04T11:31:43Z","reason":"FailedCreate","message":"Error creating: pods \"authentication-operator-6c968fdfdf-\" is forbidden: autoscaling.openshift.io/ManagementCPUsOverride the cluster does not have any nodes","type":"Warning"},{"namespace":"openshift-authentication-operator","lastTimestamp":"2025-12-04T11:35:47Z","reason":"FailedCreate","message":"Error creating: pods \"authentication-operator-6c968fdfdf-\" is forbidden: autoscaling.openshift.io/ManagementCPUsOverride the cluster does not have any nodes","type":"Warning"},{"namespace":"openshift-authentication-operator","lastTimestamp":"2025-12-04T11:35:49Z","reason":"FailedScheduling","message":"0/1 nodes are available: 1 node(s) had untolerated taint {node.kubernetes.io/not-ready: }. preemption: 0/1 nodes are available: 1 Preemption is not helpful for scheduling.","type":"Warning"},{"namespace":"openshift-authentication-operator","lastTimestamp":"2025-12-04T11:35:49Z","reason":"SuccessfulCreate","message":"Created pod: authentication-operator-6c968fdfdf-nrrfw","type":"Normal"},{"namespace":"openshift-authentication-operator","lastTimestamp":"2025-12-04T11:37:30Z","reason":"Scheduled","message":"Successfully assigned openshift-authentication-operator/authentication-operator-6c968fdfdf-nrrfw to master-0","type":"Normal"},{"namespace":"openshift-authentication-operator","lastTimestamp":"2025-12-04T11:37:33Z","reason":"AddedInterface","message":"Add eth0 [10.128.0.11/23] from ovn-kubernetes","type":"Normal"},{"namespace":"openshift-authentication-operator","lastTimestamp":"2025-12-04T11:37:33Z","reason":"Pulling","message":"Pulling image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e85850a4ae1a1e3ec2c590a4936d640882b6550124da22031c85b526afbf52df\"","type":"Normal"},{"namespace":"openshift-authentication-operator","lastTimestamp":"2025-12-04T11:37:42Z","reason":"Pulled","message":"Successfully pulled image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e85850a4ae1a1e3ec2c590a4936d640882b6550124da22031c85b526afbf52df\" in 8.745s (8.745s including waiting). Image size: 507687221 bytes.","type":"Normal"},{"namespace":"openshift-authentication-operator","lastTimestamp":"2025-12-04T11:37:42Z","reason":"Created","message":"Created container: authentication-operator","type":"Normal"},{"namespace":"openshift-authentication-operator","lastTimestamp":"2025-12-04T11:37:42Z","reason":"Started","message":"Started container authentication-operator","type":"Normal"},{"namespace":"openshift-authentication-operator","lastTimestamp":"2025-12-04T11:37:44Z","reason":"LeaderElection","message":"authentication-operator-6c968fdfdf-nrrfw_6be0e687-3ba6-4e4f-8a60-d49573d7192f became leader","type":"Normal"},{"namespace":"openshift-authentication-operator","lastTimestamp":"2025-12-04T11:37:45Z","reason":"OperatorVersionChanged","message":"clusteroperator/authentication version \"operator\" changed from \"\" to \"4.18.29\"","type":"Normal"},{"namespace":"openshift-authentication-operator","lastTimestamp":"2025-12-04T11:37:45Z","reason":"OperatorStatusChanged","message":"Status for clusteroperator/authentication changed: Degraded set to Unknown (\"\"),Progressing set to Unknown (\"\"),Available set to Unknown (\"\"),Upgradeable set to Unknown (\"\"),EvaluationConditionsDetected set to Unknown (\"\"),status.versions changed from [] to [{\"operator\" \"4.18.29\"}]","type":"Normal"},{"namespace":"openshift-authentication-operator","lastTimestamp":"2025-12-04T11:37:45Z","reason":"OperatorStatusChanged","message":"Status for clusteroperator/authentication changed: Degraded changed from Unknown to False (\"All is well\"),Available changed from Unknown to False (\"OAuthServerServiceEndpointsEndpointAccessibleControllerAvailable: endpoints \\\"oauth-openshift\\\" not found\"),Upgradeable changed from Unknown to True (\"All is well\")","type":"Normal"},{"namespace":"openshift-authentication-operator","lastTimestamp":"2025-12-04T11:37:47Z","reason":"ObserveAPIAudiences","message":"service account issuer changed from to https://kubernetes.default.svc","type":"Normal"},{"namespace":"openshift-authentication-operator","lastTimestamp":"2025-12-04T11:37:47Z","reason":"ObserveStorageUpdated","message":"Updated storage urls to https://192.168.32.10:2379","type":"Normal"},{"namespace":"openshift-authentication-operator","lastTimestamp":"2025-12-04T11:37:47Z","reason":"ObserveTLSSecurityProfile","message":"minTLSVersion changed to VersionTLS12","type":"Normal"},{"namespace":"openshift-authentication-operator","lastTimestamp":"2025-12-04T11:37:47Z","reason":"ObserveTLSSecurityProfile","message":"cipherSuites changed to [\"TLS_AES_128_GCM_SHA256\" \"TLS_AES_256_GCM_SHA384\" \"TLS_CHACHA20_POLY1305_SHA256\" \"TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256\" \"TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256\" \"TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384\" \"TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384\" \"TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305_SHA256\" \"TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305_SHA256\"]","type":"Normal"},{"namespace":"openshift-authentication-operator","lastTimestamp":"2025-12-04T11:37:47Z","reason":"ObservedConfigChanged","message":"Writing updated section (\"oauthAPIServer\") of observed config: \" map[string]any(\\n- \\tnil,\\n+ \\t{\\n+ \\t\\t\\\"apiServerArguments\\\": map[string]any{\\n+ \\t\\t\\t\\\"api-audiences\\\": []any{string(\\\"https://kubernetes.default.svc\\\")},\\n+ \\t\\t\\t\\\"cors-allowed-origins\\\": []any{string(`//127\\\\.0\\\\.0\\\\.1(:|$)`), string(\\\"//localhost(:|$)\\\")},\\n+ \\t\\t\\t\\\"etcd-servers\\\": []any{string(\\\"https://192.168.32.10:2379\\\")},\\n+ \\t\\t\\t\\\"tls-cipher-suites\\\": []any{\\n+ \\t\\t\\t\\tstring(\\\"TLS_AES_128_GCM_SHA256\\\"), string(\\\"TLS_AES_256_GCM_SHA384\\\"),\\n+ \\t\\t\\t\\tstring(\\\"TLS_CHACHA20_POLY1305_SHA256\\\"),\\n+ \\t\\t\\t\\tstring(\\\"TLS_ECDHE_ECDSA_WITH_AES_128_GCM\\\"...), ...,\\n+ \\t\\t\\t},\\n+ \\t\\t\\t\\\"tls-min-version\\\": string(\\\"VersionTLS12\\\"),\\n+ \\t\\t},\\n+ \\t},\\n )\\n\"","type":"Normal"},{"namespace":"openshift-authentication-operator","lastTimestamp":"2025-12-04T11:37:48Z","reason":"ObserveAPIServerURL","message":"loginURL changed from to https://api.sno.openstack.lab:6443","type":"Normal"},{"namespace":"openshift-authentication-operator","lastTimestamp":"2025-12-04T11:37:48Z","reason":"OperatorStatusChanged","message":"Status for clusteroperator/authentication changed: Available message changed from \"OAuthServerServiceEndpointsEndpointAccessibleControllerAvailable: endpoints \\\"oauth-openshift\\\" not found\" to \"APIServicesAvailable: endpoints \\\"api\\\" not found\\nOAuthServerServiceEndpointsEndpointAccessibleControllerAvailable: endpoints \\\"oauth-openshift\\\" not found\"","type":"Normal"},{"namespace":"openshift-authentication-operator","lastTimestamp":"2025-12-04T11:37:48Z","reason":"OperatorStatusChanged","message":"Status for clusteroperator/authentication changed: Degraded message changed from \"All is well\" to \"WebhookAuthenticatorControllerDegraded: failed to retrieve service openshift-oauth-apiserver/api: service \\\"api\\\" not found\"","type":"Normal"},{"namespace":"openshift-authentication-operator","lastTimestamp":"2025-12-04T11:37:48Z","reason":"ObserveTemplates","message":"templates changed to map[\"error\":\"/var/config/system/secrets/v4-0-config-system-ocp-branding-template/errors.html\" \"login\":\"/var/config/system/secrets/v4-0-config-system-ocp-branding-template/login.html\" \"providerSelection\":\"/var/config/system/secrets/v4-0-config-system-ocp-branding-template/providers.html\"]","type":"Normal"},{"namespace":"openshift-authentication-operator","lastTimestamp":"2025-12-04T11:37:48Z","reason":"ObserveTLSSecurityProfile","message":"minTLSVersion changed to VersionTLS12","type":"Normal"},{"namespace":"openshift-authentication-operator","lastTimestamp":"2025-12-04T11:37:48Z","reason":"ObserveTLSSecurityProfile","message":"cipherSuites changed to [\"TLS_AES_128_GCM_SHA256\" \"TLS_AES_256_GCM_SHA384\" \"TLS_CHACHA20_POLY1305_SHA256\" \"TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256\" \"TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256\" \"TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384\" \"TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384\" \"TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305_SHA256\" \"TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305_SHA256\"]","type":"Normal"},{"namespace":"openshift-authentication-operator","lastTimestamp":"2025-12-04T11:37:48Z","reason":"ObserveAuditProfile","message":"AuditProfile changed from '%!s(\u003cnil\u003e)' to 'map[audit-log-format:[json] audit-log-maxbackup:[10] audit-log-maxsize:[100] audit-log-path:[/var/log/oauth-server/audit.log] audit-policy-file:[/var/run/configmaps/audit/audit.yaml]]'","type":"Normal"},{"namespace":"openshift-authentication-operator","lastTimestamp":"2025-12-04T11:37:48Z","reason":"ObserveTokenConfig","message":"accessTokenMaxAgeSeconds changed from %!d(float64=0) to %!d(float64=86400)","type":"Normal"},{"namespace":"openshift-authentication-operator","lastTimestamp":"2025-12-04T11:37:48Z","reason":"FastControllerResync","message":"Controller \"openshift-apiserver-APIService\" resync interval is set to 10s which might lead to client request throttling","type":"Warning"},{"namespace":"openshift-authentication-operator","lastTimestamp":"2025-12-04T11:37:48Z","reason":"ObservedConfigChanged","message":"Writing updated section (\"oauthServer\") of observed config: \" map[string]any(\\n- \\tnil,\\n+ \\t{\\n+ \\t\\t\\\"corsAllowedOrigins\\\": []any{string(`//127\\\\.0\\\\.0\\\\.1(:|$)`), string(\\\"//localhost(:|$)\\\")},\\n+ \\t\\t\\\"oauthConfig\\\": map[string]any{\\n+ \\t\\t\\t\\\"assetPublicURL\\\": string(\\\"\\\"),\\n+ \\t\\t\\t\\\"loginURL\\\": string(\\\"https://api.sno.openstack.lab:6443\\\"),\\n+ \\t\\t\\t\\\"templates\\\": map[string]any{\\n+ \\t\\t\\t\\t\\\"error\\\": string(\\\"/var/config/system/secrets/v4-0-\\\"...),\\n+ \\t\\t\\t\\t\\\"login\\\": string(\\\"/var/config/system/secrets/v4-0-\\\"...),\\n+ \\t\\t\\t\\t\\\"providerSelection\\\": string(\\\"/var/config/system/secrets/v4-0-\\\"...),\\n+ \\t\\t\\t},\\n+ \\t\\t\\t\\\"tokenConfig\\\": map[string]any{\\n+ \\t\\t\\t\\t\\\"accessTokenMaxAgeSeconds\\\": float64(86400),\\n+ \\t\\t\\t\\t\\\"authorizeTokenMaxAgeSeconds\\\": float64(300),\\n+ \\t\\t\\t},\\n+ \\t\\t},\\n+ \\t\\t\\\"serverArguments\\\": map[string]any{\\n+ \\t\\t\\t\\\"audit-log-format\\\": []any{string(\\\"json\\\")},\\n+ \\t\\t\\t\\\"audit-log-maxbackup\\\": []any{string(\\\"10\\\")},\\n+ \\t\\t\\t\\\"audit-log-maxsize\\\": []any{string(\\\"100\\\")},\\n+ \\t\\t\\t\\\"audit-log-path\\\": []any{string(\\\"/var/log/oauth-server/audit.log\\\")},\\n+ \\t\\t\\t\\\"audit-policy-file\\\": []any{string(\\\"/var/run/configmaps/audit/audit.\\\"...)},\\n+ \\t\\t},\\n+ \\t\\t\\\"servingInfo\\\": map[string]any{\\n+ \\t\\t\\t\\\"cipherSuites\\\": []any{\\n+ \\t\\t\\t\\tstring(\\\"TLS_AES_128_GCM_SHA256\\\"), string(\\\"TLS_AES_256_GCM_SHA384\\\"),\\n+ \\t\\t\\t\\tstring(\\\"TLS_CHACHA20_POLY1305_SHA256\\\"),\\n+ \\t\\t\\t\\tstring(\\\"TLS_ECDHE_ECDSA_WITH_AES_128_GCM\\\"...), ...,\\n+ \\t\\t\\t},\\n+ \\t\\t\\t\\\"minTLSVersion\\\": string(\\\"VersionTLS12\\\"),\\n+ \\t\\t},\\n+ \\t\\t\\\"volumesToMount\\\": map[string]any{\\\"identityProviders\\\": string(\\\"{}\\\")},\\n+ \\t},\\n )\\n\"","type":"Normal"},{"namespace":"openshift-authentication-operator","lastTimestamp":"2025-12-04T11:37:48Z","reason":"OperatorStatusChanged","message":"Status for clusteroperator/authentication changed: Degraded message changed from \"WebhookAuthenticatorControllerDegraded: failed to retrieve service openshift-oauth-apiserver/api: service \\\"api\\\" not found\" to \"WebhookAuthenticatorControllerDegraded: failed to retrieve service openshift-oauth-apiserver/api: service \\\"api\\\" not found\\nOAuthServerConfigObservationDegraded: secret \\\"v4-0-config-system-router-certs\\\" not found\"","type":"Normal"},{"namespace":"openshift-authentication-operator","lastTimestamp":"2025-12-04T11:37:48Z","reason":"OperatorStatusChanged","message":"Status for clusteroperator/authentication changed: Progressing changed from Unknown to False (\"All is well\")","type":"Normal"},{"namespace":"openshift-authentication-operator","lastTimestamp":"2025-12-04T11:37:48Z","reason":"OperatorStatusChanged","message":"Status for clusteroperator/authentication changed: Available message changed from \"APIServicesAvailable: endpoints \\\"api\\\" not found\\nOAuthServerServiceEndpointsEndpointAccessibleControllerAvailable: endpoints \\\"oauth-openshift\\\" not found\" to \"APIServicesAvailable: endpoints \\\"api\\\" not found\\nOAuthServerServiceEndpointsEndpointAccessibleControllerAvailable: endpoints \\\"oauth-openshift\\\" not found\\nReadyIngressNodesAvailable: Authentication requires functional ingress which requires at least one schedulable and ready node. Got 0 worker nodes, 1 master nodes, 0 custom target nodes (none are schedulable or ready for ingress pods).\"","type":"Normal"},{"namespace":"openshift-authentication-operator","lastTimestamp":"2025-12-04T11:37:49Z","reason":"FastControllerResync","message":"Controller \"auditPolicyController\" resync interval is set to 10s which might lead to client request throttling","type":"Warning"},{"namespace":"openshift-authentication-operator","lastTimestamp":"2025-12-04T11:37:49Z","reason":"OperatorStatusChanged","message":"Status for clusteroperator/authentication changed: Degraded message changed from \"WebhookAuthenticatorControllerDegraded: failed to retrieve service openshift-oauth-apiserver/api: service \\\"api\\\" not found\\nOAuthServerConfigObservationDegraded: secret \\\"v4-0-config-system-router-certs\\\" not found\" to \"WebhookAuthenticatorControllerDegraded: failed to retrieve service openshift-oauth-apiserver/api: service \\\"api\\\" not found\\nOAuthServerConfigObservationDegraded: secret \\\"v4-0-config-system-router-certs\\\" not found\\nAPIServerDeploymentDegraded: waiting for .status.latestAvailableRevision to be available\\nAPIServerDeploymentDegraded: \\nAPIServerWorkloadDegraded: waiting for .status.latestAvailableRevision to be available\\nAPIServerWorkloadDegraded: \"","type":"Normal"},{"namespace":"openshift-authentication-operator","lastTimestamp":"2025-12-04T11:37:49Z","reason":"OperatorStatusChanged","message":"Status for clusteroperator/authentication changed: Degraded message changed from \"WebhookAuthenticatorControllerDegraded: failed to retrieve service openshift-oauth-apiserver/api: service \\\"api\\\" not found\\nOAuthServerConfigObservationDegraded: secret \\\"v4-0-config-system-router-certs\\\" not found\\nAPIServerDeploymentDegraded: waiting for .status.latestAvailableRevision to be available\\nAPIServerDeploymentDegraded: \\nAPIServerWorkloadDegraded: waiting for .status.latestAvailableRevision to be available\\nAPIServerWorkloadDegraded: \" to \"WebhookAuthenticatorControllerDegraded: failed to retrieve service openshift-oauth-apiserver/api: service \\\"api\\\" not found\\nOAuthServerConfigObservationDegraded: secret \\\"v4-0-config-system-router-certs\\\" not found\\nAPIServerDeploymentDegraded: waiting for .status.latestAvailableRevision to be available\\nAPIServerDeploymentDegraded: \\nAPIServerWorkloadDegraded: waiting for .status.latestAvailableRevision to be available\\nAPIServerWorkloadDegraded: \\nRevisionControllerDegraded: configmap \\\"audit\\\" not found\"","type":"Normal"},{"namespace":"openshift-authentication-operator","lastTimestamp":"2025-12-04T11:37:49Z","reason":"OperatorStatusChanged","message":"Status for clusteroperator/authentication changed: Degraded message changed from \"WebhookAuthenticatorControllerDegraded: failed to retrieve service openshift-oauth-apiserver/api: service \\\"api\\\" not found\\nOAuthServerConfigObservationDegraded: secret \\\"v4-0-config-system-router-certs\\\" not found\\nAPIServerDeploymentDegraded: waiting for .status.latestAvailableRevision to be available\\nAPIServerDeploymentDegraded: \\nAPIServerWorkloadDegraded: waiting for .status.latestAvailableRevision to be available\\nAPIServerWorkloadDegraded: \\nRevisionControllerDegraded: configmap \\\"audit\\\" not found\" to \"WebhookAuthenticatorControllerDegraded: failed to retrieve service openshift-oauth-apiserver/api: service \\\"api\\\" not found\\nOAuthServerConfigObservationDegraded: secret \\\"v4-0-config-system-router-certs\\\" not found\\nAPIServerDeploymentDegraded: waiting for .status.latestAvailableRevision to be available\\nAPIServerDeploymentDegraded: \\nAPIServerWorkloadDegraded: waiting for .status.latestAvailableRevision to be available\\nAPIServerWorkloadDegraded: \\nRevisionControllerDegraded: configmap \\\"audit\\\" not found\\nOAuthServiceDegraded: Unable to get oauth server service: service \\\"oauth-openshift\\\" not found\"","type":"Normal"},{"namespace":"openshift-authentication-operator","lastTimestamp":"2025-12-04T11:37:49Z","reason":"OperatorStatusChanged","message":"Status for clusteroperator/authentication changed: Available message changed from \"APIServicesAvailable: endpoints \\\"api\\\" not found\\nOAuthServerServiceEndpointsEndpointAccessibleControllerAvailable: endpoints \\\"oauth-openshift\\\" not found\\nReadyIngressNodesAvailable: Authentication requires functional ingress which requires at least one schedulable and ready node. Got 0 worker nodes, 1 master nodes, 0 custom target nodes (none are schedulable or ready for ingress pods).\" to \"APIServicesAvailable: endpoints \\\"api\\\" not found\\nOAuthServerServiceEndpointAccessibleControllerAvailable: service \\\"oauth-openshift\\\" not found\\nOAuthServerServiceEndpointsEndpointAccessibleControllerAvailable: endpoints \\\"oauth-openshift\\\" not found\\nReadyIngressNodesAvailable: Authentication requires functional ingress which requires at least one schedulable and ready node. Got 0 worker nodes, 1 master nodes, 0 custom target nodes (none are schedulable or ready for ingress pods).\"","type":"Normal"},{"namespace":"openshift-authentication-operator","lastTimestamp":"2025-12-04T11:37:50Z","reason":"NoValidCertificateFound","message":"No valid client certificate for OpenShiftAuthenticatorCertRequester is found: unable to parse certificate: data does not contain any valid RSA or ECDSA certificates","type":"Normal"},{"namespace":"openshift-authentication-operator","lastTimestamp":"2025-12-04T11:37:50Z","reason":"OperatorStatusChanged","message":"Status for clusteroperator/authentication changed: Degraded message changed from \"WebhookAuthenticatorControllerDegraded: failed to retrieve service openshift-oauth-apiserver/api: service \\\"api\\\" not found\\nOAuthServerConfigObservationDegraded: secret \\\"v4-0-config-system-router-certs\\\" not found\\nAPIServerDeploymentDegraded: waiting for .status.latestAvailableRevision to be available\\nAPIServerDeploymentDegraded: \\nAPIServerWorkloadDegraded: waiting for .status.latestAvailableRevision to be available\\nAPIServerWorkloadDegraded: \\nRevisionControllerDegraded: configmap \\\"audit\\\" not found\\nOAuthServiceDegraded: Unable to get oauth server service: service \\\"oauth-openshift\\\" not found\" to \"WebhookAuthenticatorControllerDegraded: failed to retrieve service openshift-oauth-apiserver/api: service \\\"api\\\" not found\\nOAuthServerConfigObservationDegraded: secret \\\"v4-0-config-system-router-certs\\\" not found\\nAPIServerDeploymentDegraded: waiting for .status.latestAvailableRevision to be available\\nAPIServerDeploymentDegraded: \\nAPIServerWorkloadDegraded: waiting for .status.latestAvailableRevision to be available\\nAPIServerWorkloadDegraded: \\nRevisionControllerDegraded: configmap \\\"audit\\\" not found\\nOAuthServiceDegraded: Unable to get oauth server service: service \\\"oauth-openshift\\\" not found\\nIngressStateEndpointsDegraded: No endpoints found for oauth-server\"","type":"Normal"},{"namespace":"openshift-authentication-operator","lastTimestamp":"2025-12-04T11:37:50Z","reason":"CSRApproval","message":"The CSR \"system:openshift:openshift-authenticator-pwb88\" has been approved","type":"Normal"},{"namespace":"openshift-authentication-operator","lastTimestamp":"2025-12-04T11:37:50Z","reason":"CSRCreated","message":"A csr \"system:openshift:openshift-authenticator-pwb88\" is created for OpenShiftAuthenticatorCertRequester","type":"Normal"},{"namespace":"openshift-authentication-operator","lastTimestamp":"2025-12-04T11:37:51Z","reason":"ConfigMapCreateFailed","message":"Failed to create ConfigMap/etcd-serving-ca -n openshift-oauth-apiserver: namespaces \"openshift-oauth-apiserver\" not found","type":"Warning"},{"namespace":"openshift-authentication-operator","lastTimestamp":"2025-12-04T11:37:52Z","reason":"ConfigMapCreateFailed","message":"Failed to create ConfigMap/audit -n openshift-authentication: namespaces \"openshift-authentication\" not found","type":"Warning"},{"namespace":"openshift-authentication-operator","lastTimestamp":"2025-12-04T11:37:52Z","reason":"OperatorStatusChanged","message":"Status for clusteroperator/authentication changed: Degraded message changed from \"WebhookAuthenticatorControllerDegraded: failed to retrieve service openshift-oauth-apiserver/api: service \\\"api\\\" not found\\nOAuthServerConfigObservationDegraded: secret \\\"v4-0-config-system-router-certs\\\" not found\\nAPIServerDeploymentDegraded: waiting for .status.latestAvailableRevision to be available\\nAPIServerDeploymentDegraded: \\nAPIServerWorkloadDegraded: waiting for .status.latestAvailableRevision to be available\\nAPIServerWorkloadDegraded: \\nRevisionControllerDegraded: configmap \\\"audit\\\" not found\\nOAuthServiceDegraded: Unable to get oauth server service: service \\\"oauth-openshift\\\" not found\\nIngressStateEndpointsDegraded: No endpoints found for oauth-server\" to \"WebhookAuthenticatorControllerDegraded: failed to retrieve service openshift-oauth-apiserver/api: service \\\"api\\\" not found\\nOAuthServerConfigObservationDegraded: secret \\\"v4-0-config-system-router-certs\\\" not found\\nAPIServerDeploymentDegraded: waiting for .status.latestAvailableRevision to be available\\nAPIServerDeploymentDegraded: \\nAPIServerWorkloadDegraded: waiting for .status.latestAvailableRevision to be available\\nAPIServerWorkloadDegraded: \\nRevisionControllerDegraded: configmap \\\"audit\\\" not found\\nOAuthServiceDegraded: Unable to get oauth server service: service \\\"oauth-openshift\\\" not found\\nIngressStateEndpointsDegraded: No endpoints found for oauth-server\\nRouterCertsDegraded: neither the custom secret/v4-0-config-system-router-certs -n openshift-authentication or default secret/v4-0-config-system-custom-router-certs -n openshift-authentication could be retrieved: secret \\\"v4-0-config-system-router-certs\\\" not found\"","type":"Normal"},{"namespace":"openshift-authentication-operator","lastTimestamp":"2025-12-04T11:37:52Z","reason":"NamespaceCreated","message":"Created Namespace/openshift-oauth-apiserver because it was missing","type":"Normal"},{"namespace":"openshift-authentication-operator","lastTimestamp":"2025-12-04T11:37:52Z","reason":"ClusterRoleBindingCreated","message":"Created ClusterRoleBinding.rbac.authorization.k8s.io/system:openshift:oauth-apiserver because it was missing","type":"Normal"},{"namespace":"openshift-authentication-operator","lastTimestamp":"2025-12-04T11:37:53Z","reason":"OperatorStatusChanged","message":"Status for clusteroperator/authentication changed: Degraded message changed from \"WebhookAuthenticatorControllerDegraded: failed to retrieve service openshift-oauth-apiserver/api: service \\\"api\\\" not found\\nOAuthServerConfigObservationDegraded: secret \\\"v4-0-config-system-router-certs\\\" not found\\nAPIServerDeploymentDegraded: waiting for .status.latestAvailableRevision to be available\\nAPIServerDeploymentDegraded: \\nAPIServerWorkloadDegraded: waiting for .status.latestAvailableRevision to be available\\nAPIServerWorkloadDegraded: \\nRevisionControllerDegraded: configmap \\\"audit\\\" not found\\nOAuthServiceDegraded: Unable to get oauth server service: service \\\"oauth-openshift\\\" not found\\nIngressStateEndpointsDegraded: No endpoints found for oauth-server\\nRouterCertsDegraded: neither the custom secret/v4-0-config-system-router-certs -n openshift-authentication or default secret/v4-0-config-system-custom-router-certs -n openshift-authentication could be retrieved: secret \\\"v4-0-config-system-router-certs\\\" not found\" to \"WebhookAuthenticatorControllerDegraded: failed to retrieve service openshift-oauth-apiserver/api: service \\\"api\\\" not found\\nOAuthServerConfigObservationDegraded: secret \\\"v4-0-config-system-router-certs\\\" not found\\nAPIServerDeploymentDegraded: waiting for .status.latestAvailableRevision to be available\\nAPIServerDeploymentDegraded: \\nAPIServerWorkloadDegraded: waiting for .status.latestAvailableRevision to be available\\nAPIServerWorkloadDegraded: \\nRevisionControllerDegraded: configmap \\\"audit\\\" not found\\nOAuthServiceDegraded: Unable to get oauth server service: service \\\"oauth-openshift\\\" not found\\nIngressStateEndpointsDegraded: No endpoints found for oauth-server\\nRouterCertsDegraded: neither the custom secret/v4-0-config-system-router-certs -n openshift-authentication or default secret/v4-0-config-system-custom-router-certs -n openshift-authentication could be retrieved: secret \\\"v4-0-config-system-router-certs\\\" not found\\nResourceSyncControllerDegraded: namespaces \\\"openshift-oauth-apiserver\\\" not found\"","type":"Normal"},{"namespace":"openshift-authentication-operator","lastTimestamp":"2025-12-04T11:37:54Z","reason":"ConfigMapCreated","message":"Created ConfigMap/audit -n openshift-oauth-apiserver because it was missing","type":"Normal"},{"namespace":"openshift-authentication-operator","lastTimestamp":"2025-12-04T11:37:54Z","reason":"StartingNewRevision","message":"new revision 1 triggered by \"configmap \\\"audit-0\\\" not found\"","type":"Normal"},{"namespace":"openshift-authentication-operator","lastTimestamp":"2025-12-04T11:37:55Z","reason":"ClientCertificateCreated","message":"A new client certificate for OpenShiftAuthenticatorCertRequester is available","type":"Normal"},{"namespace":"openshift-authentication-operator","lastTimestamp":"2025-12-04T11:37:56Z","reason":"ConfigMapCreated","message":"Created ConfigMap/etcd-serving-ca -n openshift-oauth-apiserver because it was missing","type":"Normal"},{"namespace":"openshift-authentication-operator","lastTimestamp":"2025-12-04T11:37:57Z","reason":"NamespaceCreated","message":"Created Namespace/openshift-authentication because it was missing","type":"Normal"},{"namespace":"openshift-authentication-operator","lastTimestamp":"2025-12-04T11:37:57Z","reason":"ClusterRoleBindingCreated","message":"Created ClusterRoleBinding.rbac.authorization.k8s.io/system:openshift:openshift-authentication because it was missing","type":"Normal"},{"namespace":"openshift-authentication-operator","lastTimestamp":"2025-12-04T11:37:58Z","reason":"ServiceCreated","message":"Created Service/api -n openshift-oauth-apiserver because it was missing","type":"Normal"},{"namespace":"openshift-authentication-operator","lastTimestamp":"2025-12-04T11:37:58Z","reason":"OperatorStatusChanged","message":"Status for clusteroperator/authentication changed: Progressing message changed from \"All is well\" to \"AuthenticatorCertKeyProgressing: All is well\"","type":"Normal"},{"namespace":"openshift-authentication-operator","lastTimestamp":"2025-12-04T11:37:58Z","reason":"OperatorStatusChanged","message":"Status for clusteroperator/authentication changed: Degraded message changed from \"WebhookAuthenticatorControllerDegraded: failed to retrieve service openshift-oauth-apiserver/api: service \\\"api\\\" not found\\nOAuthServerConfigObservationDegraded: secret \\\"v4-0-config-system-router-certs\\\" not found\\nAPIServerDeploymentDegraded: waiting for .status.latestAvailableRevision to be available\\nAPIServerDeploymentDegraded: \\nAPIServerWorkloadDegraded: waiting for .status.latestAvailableRevision to be available\\nAPIServerWorkloadDegraded: \\nRevisionControllerDegraded: configmap \\\"audit\\\" not found\\nOAuthServiceDegraded: Unable to get oauth server service: service \\\"oauth-openshift\\\" not found\\nIngressStateEndpointsDegraded: No endpoints found for oauth-server\\nRouterCertsDegraded: neither the custom secret/v4-0-config-system-router-certs -n openshift-authentication or default secret/v4-0-config-system-custom-router-certs -n openshift-authentication could be retrieved: secret \\\"v4-0-config-system-router-certs\\\" not found\\nResourceSyncControllerDegraded: namespaces \\\"openshift-oauth-apiserver\\\" not found\" to \"WebhookAuthenticatorControllerDegraded: failed to read service-ca crt bundle: open /var/run/configmaps/service-ca-bundle/service-ca.crt: no such file or directory\\nOAuthServerConfigObservationDegraded: secret \\\"v4-0-config-system-router-certs\\\" not found\\nAPIServerDeploymentDegraded: waiting for .status.latestAvailableRevision to be available\\nAPIServerDeploymentDegraded: \\nAPIServerWorkloadDegraded: waiting for .status.latestAvailableRevision to be available\\nAPIServerWorkloadDegraded: \\nRevisionControllerDegraded: configmap \\\"audit\\\" not found\\nOAuthServiceDegraded: Unable to get oauth server service: service \\\"oauth-openshift\\\" not found\\nIngressStateEndpointsDegraded: No endpoints found for oauth-server\\nRouterCertsDegraded: neither the custom secret/v4-0-config-system-router-certs -n openshift-authentication or default secret/v4-0-config-system-custom-router-certs -n openshift-authentication could be retrieved: secret \\\"v4-0-config-system-router-certs\\\" not found\\nResourceSyncControllerDegraded: namespaces \\\"openshift-oauth-apiserver\\\" not found\"","type":"Normal"},{"namespace":"openshift-authentication-operator","lastTimestamp":"2025-12-04T11:37:58Z","reason":"OperatorStatusChanged","message":"Status for clusteroperator/authentication changed: Available message changed from \"APIServicesAvailable: endpoints \\\"api\\\" not found\\nOAuthServerServiceEndpointAccessibleControllerAvailable: service \\\"oauth-openshift\\\" not found\\nOAuthServerServiceEndpointsEndpointAccessibleControllerAvailable: endpoints \\\"oauth-openshift\\\" not found\\nReadyIngressNodesAvailable: Authentication requires functional ingress which requires at least one schedulable and ready node. Got 0 worker nodes, 1 master nodes, 0 custom target nodes (none are schedulable or ready for ingress pods).\" to \"APIServicesAvailable: PreconditionNotReady\\nOAuthServerServiceEndpointAccessibleControllerAvailable: service \\\"oauth-openshift\\\" not found\\nOAuthServerServiceEndpointsEndpointAccessibleControllerAvailable: endpoints \\\"oauth-openshift\\\" not found\\nReadyIngressNodesAvailable: Authentication requires functional ingress which requires at least one schedulable and ready node. Got 0 worker nodes, 1 master nodes, 0 custom target nodes (none are schedulable or ready for ingress pods).\"","type":"Normal"},{"namespace":"openshift-authentication-operator","lastTimestamp":"2025-12-04T11:37:59Z","reason":"SecretCreated","message":"Created Secret/etcd-client -n openshift-oauth-apiserver because it was missing","type":"Normal"},{"namespace":"openshift-authentication-operator","lastTimestamp":"2025-12-04T11:37:59Z","reason":"OperatorStatusChanged","message":"Status for clusteroperator/authentication changed: Degraded message changed from \"WebhookAuthenticatorControllerDegraded: failed to read service-ca crt bundle: open /var/run/configmaps/service-ca-bundle/service-ca.crt: no such file or directory\\nOAuthServerConfigObservationDegraded: secret \\\"v4-0-config-system-router-certs\\\" not found\\nAPIServerDeploymentDegraded: waiting for .status.latestAvailableRevision to be available\\nAPIServerDeploymentDegraded: \\nAPIServerWorkloadDegraded: waiting for .status.latestAvailableRevision to be available\\nAPIServerWorkloadDegraded: \\nRevisionControllerDegraded: configmap \\\"audit\\\" not found\\nOAuthServiceDegraded: Unable to get oauth server service: service \\\"oauth-openshift\\\" not found\\nIngressStateEndpointsDegraded: No endpoints found for oauth-server\\nRouterCertsDegraded: neither the custom secret/v4-0-config-system-router-certs -n openshift-authentication or default secret/v4-0-config-system-custom-router-certs -n openshift-authentication could be retrieved: secret \\\"v4-0-config-system-router-certs\\\" not found\\nResourceSyncControllerDegraded: namespaces \\\"openshift-oauth-apiserver\\\" not found\" to \"WebhookAuthenticatorControllerDegraded: failed to read service-ca crt bundle: open /var/run/configmaps/service-ca-bundle/service-ca.crt: no such file or directory\\nOAuthServerConfigObservationDegraded: secret \\\"v4-0-config-system-router-certs\\\" not found\\nAPIServerDeploymentDegraded: waiting for .status.latestAvailableRevision to be available\\nAPIServerDeploymentDegraded: \\nAPIServerWorkloadDegraded: waiting for .status.latestAvailableRevision to be available\\nAPIServerWorkloadDegraded: \\nRevisionControllerDegraded: configmap \\\"audit\\\" not found\\nOAuthServiceDegraded: Unable to get oauth server service: service \\\"oauth-openshift\\\" not found\\nIngressStateEndpointsDegraded: No endpoints found for oauth-server\\nRouterCertsDegraded: neither the custom secret/v4-0-config-system-router-certs -n openshift-authentication or default secret/v4-0-config-system-custom-router-certs -n openshift-authentication could be retrieved: secret \\\"v4-0-config-system-router-certs\\\" not found\"","type":"Normal"},{"namespace":"openshift-authentication-operator","lastTimestamp":"2025-12-04T11:37:59Z","reason":"ConfigMapCreated","message":"Created ConfigMap/audit-1 -n openshift-oauth-apiserver because it was missing","type":"Normal"},{"namespace":"openshift-authentication-operator","lastTimestamp":"2025-12-04T11:38:00Z","reason":"ConfigMapCreated","message":"Created ConfigMap/v4-0-config-system-trusted-ca-bundle -n openshift-authentication because it was missing","type":"Normal"},{"namespace":"openshift-authentication-operator","lastTimestamp":"2025-12-04T11:38:02Z","reason":"SecretCreated","message":"Created Secret/v4-0-config-system-ocp-branding-template -n openshift-authentication because it was missing","type":"Normal"},{"namespace":"openshift-authentication-operator","lastTimestamp":"2025-12-04T11:38:03Z","reason":"ServiceAccountCreated","message":"Created ServiceAccount/oauth-apiserver-sa -n openshift-oauth-apiserver because it was missing","type":"Normal"},{"namespace":"openshift-authentication-operator","lastTimestamp":"2025-12-04T11:38:03Z","reason":"ClusterRoleBindingCreated","message":"Created ClusterRoleBinding.rbac.authorization.k8s.io/system:openshift:useroauthaccesstoken-manager because it was missing","type":"Normal"},{"namespace":"openshift-authentication-operator","lastTimestamp":"2025-12-04T11:38:03Z","reason":"ClusterRoleCreated","message":"Created ClusterRole.rbac.authorization.k8s.io/system:openshift:useroauthaccesstoken-manager because it was missing","type":"Normal"},{"namespace":"openshift-authentication-operator","lastTimestamp":"2025-12-04T11:38:04Z","reason":"RevisionTriggered","message":"new revision 1 triggered by \"configmap \\\"audit-0\\\" not found\"","type":"Normal"},{"namespace":"openshift-authentication-operator","lastTimestamp":"2025-12-04T11:38:04Z","reason":"OperatorStatusChanged","message":"Status for clusteroperator/authentication changed: Degraded message changed from \"WebhookAuthenticatorControllerDegraded: failed to read service-ca crt bundle: open /var/run/configmaps/service-ca-bundle/service-ca.crt: no such file or directory\\nOAuthServerConfigObservationDegraded: secret \\\"v4-0-config-system-router-certs\\\" not found\\nAPIServerDeploymentDegraded: waiting for .status.latestAvailableRevision to be available\\nAPIServerDeploymentDegraded: \\nAPIServerWorkloadDegraded: waiting for .status.latestAvailableRevision to be available\\nAPIServerWorkloadDegraded: \\nRevisionControllerDegraded: configmap \\\"audit\\\" not found\\nOAuthServiceDegraded: Unable to get oauth server service: service \\\"oauth-openshift\\\" not found\\nIngressStateEndpointsDegraded: No endpoints found for oauth-server\\nRouterCertsDegraded: neither the custom secret/v4-0-config-system-router-certs -n openshift-authentication or default secret/v4-0-config-system-custom-router-certs -n openshift-authentication could be retrieved: secret \\\"v4-0-config-system-router-certs\\\" not found\" to \"WebhookAuthenticatorControllerDegraded: failed to read service-ca crt bundle: open /var/run/configmaps/service-ca-bundle/service-ca.crt: no such file or directory\\nOAuthServerConfigObservationDegraded: secret \\\"v4-0-config-system-router-certs\\\" not found\\nAPIServerDeploymentDegraded: waiting for .status.latestAvailableRevision to be available\\nAPIServerDeploymentDegraded: \\nAPIServerWorkloadDegraded: waiting for .status.latestAvailableRevision to be available\\nAPIServerWorkloadDegraded: \\nOAuthServiceDegraded: Unable to get oauth server service: service \\\"oauth-openshift\\\" not found\\nIngressStateEndpointsDegraded: No endpoints found for oauth-server\\nRouterCertsDegraded: neither the custom secret/v4-0-config-system-router-certs -n openshift-authentication or default secret/v4-0-config-system-custom-router-certs -n openshift-authentication could be retrieved: secret \\\"v4-0-config-system-router-certs\\\" not found\"","type":"Normal"},{"namespace":"openshift-authentication-operator","lastTimestamp":"2025-12-04T11:38:08Z","reason":"ServiceAccountCreated","message":"Created ServiceAccount/oauth-openshift -n openshift-authentication because it was missing","type":"Normal"},{"namespace":"openshift-authentication-operator","lastTimestamp":"2025-12-04T11:38:12Z","reason":"ServiceCreated","message":"Created Service/oauth-openshift -n openshift-authentication because it was missing","type":"Normal"},{"namespace":"openshift-authentication-operator","lastTimestamp":"2025-12-04T11:38:14Z","reason":"DeploymentCreated","message":"Created Deployment.apps/apiserver -n openshift-oauth-apiserver because it was missing","type":"Normal"},{"namespace":"openshift-authentication-operator","lastTimestamp":"2025-12-04T11:38:18Z","reason":"RoleCreated","message":"Created Role.rbac.authorization.k8s.io/system:openshift:oauth-servercert-trust -n openshift-config-managed because it was missing","type":"Normal"},{"namespace":"openshift-authentication-operator","lastTimestamp":"2025-12-04T11:38:18Z","reason":"OperatorStatusChanged","message":"Status for clusteroperator/authentication changed: Degraded message changed from \"WebhookAuthenticatorControllerDegraded: failed to read service-ca crt bundle: open /var/run/configmaps/service-ca-bundle/service-ca.crt: no such file or directory\\nOAuthServerConfigObservationDegraded: secret \\\"v4-0-config-system-router-certs\\\" not found\\nAPIServerDeploymentDegraded: waiting for .status.latestAvailableRevision to be available\\nAPIServerDeploymentDegraded: \\nAPIServerWorkloadDegraded: waiting for .status.latestAvailableRevision to be available\\nAPIServerWorkloadDegraded: \\nOAuthServiceDegraded: Unable to get oauth server service: service \\\"oauth-openshift\\\" not found\\nIngressStateEndpointsDegraded: No endpoints found for oauth-server\\nRouterCertsDegraded: neither the custom secret/v4-0-config-system-router-certs -n openshift-authentication or default secret/v4-0-config-system-custom-router-certs -n openshift-authentication could be retrieved: secret \\\"v4-0-config-system-router-certs\\\" not found\" to \"WebhookAuthenticatorControllerDegraded: failed to read service-ca crt bundle: open /var/run/configmaps/service-ca-bundle/service-ca.crt: no such file or directory\\nOAuthServerConfigObservationDegraded: secret \\\"v4-0-config-system-router-certs\\\" not found\\nAPIServerDeploymentDegraded: waiting for .status.latestAvailableRevision to be available\\nAPIServerDeploymentDegraded: \\nAPIServerWorkloadDegraded: waiting for .status.latestAvailableRevision to be available\\nAPIServerWorkloadDegraded: \\nOAuthServiceDegraded: Unable to get oauth server service: service \\\"oauth-openshift\\\" not found\\nIngressStateEndpointsDegraded: No endpoints found for oauth-server\\nOAuthServerServiceEndpointAccessibleControllerDegraded: configmap \\\"v4-0-config-system-service-ca\\\" not found\\nRouterCertsDegraded: neither the custom secret/v4-0-config-system-router-certs -n openshift-authentication or default secret/v4-0-config-system-custom-router-certs -n openshift-authentication could be retrieved: secret \\\"v4-0-config-system-router-certs\\\" not found\"","type":"Normal"},{"namespace":"openshift-authentication-operator","lastTimestamp":"2025-12-04T11:38:18Z","reason":"OperatorStatusChanged","message":"Status for clusteroperator/authentication changed: Degraded message changed from \"WebhookAuthenticatorControllerDegraded: failed to read service-ca crt bundle: open /var/run/configmaps/service-ca-bundle/service-ca.crt: no such file or directory\\nOAuthServerConfigObservationDegraded: secret \\\"v4-0-config-system-router-certs\\\" not found\\nAPIServerDeploymentDegraded: waiting for .status.latestAvailableRevision to be available\\nAPIServerDeploymentDegraded: \\nAPIServerWorkloadDegraded: waiting for .status.latestAvailableRevision to be available\\nAPIServerWorkloadDegraded: \\nOAuthServiceDegraded: Unable to get oauth server service: service \\\"oauth-openshift\\\" not found\\nIngressStateEndpointsDegraded: No endpoints found for oauth-server\\nOAuthServerServiceEndpointAccessibleControllerDegraded: configmap \\\"v4-0-config-system-service-ca\\\" not found\\nRouterCertsDegraded: neither the custom secret/v4-0-config-system-router-certs -n openshift-authentication or default secret/v4-0-config-system-custom-router-certs -n openshift-authentication could be retrieved: secret \\\"v4-0-config-system-router-certs\\\" not found\" to \"WebhookAuthenticatorControllerDegraded: failed to read service-ca crt bundle: open /var/run/configmaps/service-ca-bundle/service-ca.crt: no such file or directory\\nOAuthServerConfigObservationDegraded: secret \\\"v4-0-config-system-router-certs\\\" not found\\nOAuthServiceDegraded: Unable to get oauth server service: service \\\"oauth-openshift\\\" not found\\nIngressStateEndpointsDegraded: No endpoints found for oauth-server\\nOAuthServerServiceEndpointAccessibleControllerDegraded: \\\"service-ca.crt\\\" key of the \\\"openshift-authentication/v4-0-config-system-service-ca\\\" CM is empty\\nRouterCertsDegraded: neither the custom secret/v4-0-config-system-router-certs -n openshift-authentication or default secret/v4-0-config-system-custom-router-certs -n openshift-authentication could be retrieved: secret \\\"v4-0-config-system-router-certs\\\" not found\",Progressing changed from False to True (\"APIServerDeploymentProgressing: deployment/apiserver.openshift-oauth-apiserver: observed generation is 0, desired generation is 1.\"),Available message changed from \"APIServicesAvailable: PreconditionNotReady\\nOAuthServerServiceEndpointAccessibleControllerAvailable: service \\\"oauth-openshift\\\" not found\\nOAuthServerServiceEndpointsEndpointAccessibleControllerAvailable: endpoints \\\"oauth-openshift\\\" not found\\nReadyIngressNodesAvailable: Authentication requires functional ingress which requires at least one schedulable and ready node. Got 0 worker nodes, 1 master nodes, 0 custom target nodes (none are schedulable or ready for ingress pods).\" to \"APIServerDeploymentAvailable: no apiserver.openshift-oauth-apiserver pods available on any node.\\nAPIServicesAvailable: PreconditionNotReady\\nOAuthServerServiceEndpointAccessibleControllerAvailable: service \\\"oauth-openshift\\\" not found\\nOAuthServerServiceEndpointsEndpointAccessibleControllerAvailable: endpoints \\\"oauth-openshift\\\" not found\\nReadyIngressNodesAvailable: Authentication requires functional ingress which requires at least one schedulable and ready node. Got 0 worker nodes, 1 master nodes, 0 custom target nodes (none are schedulable or ready for ingress pods).\"","type":"Normal"},{"namespace":"openshift-authentication-operator","lastTimestamp":"2025-12-04T11:38:18Z","reason":"RoleBindingCreated","message":"Created RoleBinding.rbac.authorization.k8s.io/system:openshift:oauth-servercert-trust -n openshift-config-managed because it was missing","type":"Normal"},{"namespace":"openshift-authentication-operator","lastTimestamp":"2025-12-04T11:38:18Z","reason":"OperatorStatusChanged","message":"Status for clusteroperator/authentication changed: Degraded message changed from \"WebhookAuthenticatorControllerDegraded: failed to read service-ca crt bundle: open /var/run/configmaps/service-ca-bundle/service-ca.crt: no such file or directory\\nOAuthServerConfigObservationDegraded: secret \\\"v4-0-config-system-router-certs\\\" not found\\nOAuthServiceDegraded: Unable to get oauth server service: service \\\"oauth-openshift\\\" not found\\nIngressStateEndpointsDegraded: No endpoints found for oauth-server\\nOAuthServerServiceEndpointAccessibleControllerDegraded: \\\"service-ca.crt\\\" key of the \\\"openshift-authentication/v4-0-config-system-service-ca\\\" CM is empty\\nRouterCertsDegraded: neither the custom secret/v4-0-config-system-router-certs -n openshift-authentication or default secret/v4-0-config-system-custom-router-certs -n openshift-authentication could be retrieved: secret \\\"v4-0-config-system-router-certs\\\" not found\" to \"OAuthServerServiceEndpointsEndpointAccessibleControllerDegraded: oauth service endpoints are not ready\\nWebhookAuthenticatorControllerDegraded: failed to read service-ca crt bundle: open /var/run/configmaps/service-ca-bundle/service-ca.crt: no such file or directory\\nOAuthServerConfigObservationDegraded: secret \\\"v4-0-config-system-router-certs\\\" not found\\nSystemServiceCAConfigDegraded: Config \\\"\\\" has no service CA data\\nIngressStateEndpointsDegraded: No endpoints found for oauth-server\\nOAuthServerServiceEndpointAccessibleControllerDegraded: \\\"service-ca.crt\\\" key of the \\\"openshift-authentication/v4-0-config-system-service-ca\\\" CM is empty\\nRouterCertsDegraded: neither the custom secret/v4-0-config-system-router-certs -n openshift-authentication or default secret/v4-0-config-system-custom-router-certs -n openshift-authentication could be retrieved: secret \\\"v4-0-config-system-router-certs\\\" not found\"","type":"Normal"},{"namespace":"openshift-authentication-operator","lastTimestamp":"2025-12-04T11:38:19Z","reason":"OperatorStatusChanged","message":"Status for clusteroperator/authentication changed: Available message changed from \"APIServerDeploymentAvailable: no apiserver.openshift-oauth-apiserver pods available on any node.\\nAPIServicesAvailable: PreconditionNotReady\\nOAuthServerServiceEndpointAccessibleControllerAvailable: service \\\"oauth-openshift\\\" not found\\nOAuthServerServiceEndpointsEndpointAccessibleControllerAvailable: endpoints \\\"oauth-openshift\\\" not found\\nReadyIngressNodesAvailable: Authentication requires functional ingress which requires at least one schedulable and ready node. Got 0 worker nodes, 1 master nodes, 0 custom target nodes (none are schedulable or ready for ingress pods).\" to \"APIServerDeploymentAvailable: no apiserver.openshift-oauth-apiserver pods available on any node.\\nAPIServicesAvailable: PreconditionNotReady\\nOAuthServerServiceEndpointAccessibleControllerAvailable: Get \\\"https://172.30.202.245:443/healthz\\\": dial tcp 172.30.202.245:443: connect: connection refused\\nOAuthServerServiceEndpointsEndpointAccessibleControllerAvailable: endpoints \\\"oauth-openshift\\\" not found\\nReadyIngressNodesAvailable: Authentication requires functional ingress which requires at least one schedulable and ready node. Got 0 worker nodes, 1 master nodes, 0 custom target nodes (none are schedulable or ready for ingress pods).\"","type":"Normal"},{"namespace":"openshift-authentication-operator","lastTimestamp":"2025-12-04T11:38:19Z","reason":"OperatorStatusChanged","message":"Status for clusteroperator/authentication changed: Degraded message changed from \"OAuthServerServiceEndpointsEndpointAccessibleControllerDegraded: oauth service endpoints are not ready\\nWebhookAuthenticatorControllerDegraded: failed to read service-ca crt bundle: open /var/run/configmaps/service-ca-bundle/service-ca.crt: no such file or directory\\nOAuthServerConfigObservationDegraded: secret \\\"v4-0-config-system-router-certs\\\" not found\\nSystemServiceCAConfigDegraded: Config \\\"\\\" has no service CA data\\nIngressStateEndpointsDegraded: No endpoints found for oauth-server\\nOAuthServerServiceEndpointAccessibleControllerDegraded: \\\"service-ca.crt\\\" key of the \\\"openshift-authentication/v4-0-config-system-service-ca\\\" CM is empty\\nRouterCertsDegraded: neither the custom secret/v4-0-config-system-router-certs -n openshift-authentication or default secret/v4-0-config-system-custom-router-certs -n openshift-authentication could be retrieved: secret \\\"v4-0-config-system-router-certs\\\" not found\" to \"OAuthServerServiceEndpointsEndpointAccessibleControllerDegraded: oauth service endpoints are not ready\\nWebhookAuthenticatorControllerDegraded: failed to read service-ca crt bundle: open /var/run/configmaps/service-ca-bundle/service-ca.crt: no such file or directory\\nOAuthServerConfigObservationDegraded: secret \\\"v4-0-config-system-router-certs\\\" not found\\nSystemServiceCAConfigDegraded: Config \\\"\\\" has no service CA data\\nIngressStateEndpointsDegraded: No endpoints found for oauth-server\\nOAuthServerServiceEndpointAccessibleControllerDegraded: Get \\\"https://172.30.202.245:443/healthz\\\": dial tcp 172.30.202.245:443: connect: connection refused\\nRouterCertsDegraded: neither the custom secret/v4-0-config-system-router-certs -n openshift-authentication or default secret/v4-0-config-system-custom-router-certs -n openshift-authentication could be retrieved: secret \\\"v4-0-config-system-router-certs\\\" not found\\nOpenshiftAuthenticationStaticResourcesDegraded: \\\"oauth-openshift/audit-policy.yaml\\\" (string): namespaces \\\"openshift-authentication\\\" not found\\nOpenshiftAuthenticationStaticResourcesDegraded: \"","type":"Normal"},{"namespace":"openshift-authentication-operator","lastTimestamp":"2025-12-04T11:38:19Z","reason":"OperatorStatusChanged","message":"Status for clusteroperator/authentication changed: Degraded message changed from \"OAuthServerServiceEndpointsEndpointAccessibleControllerDegraded: oauth service endpoints are not ready\\nWebhookAuthenticatorControllerDegraded: failed to read service-ca crt bundle: open /var/run/configmaps/service-ca-bundle/service-ca.crt: no such file or directory\\nOAuthServerConfigObservationDegraded: secret \\\"v4-0-config-system-router-certs\\\" not found\\nSystemServiceCAConfigDegraded: Config \\\"\\\" has no service CA data\\nIngressStateEndpointsDegraded: No endpoints found for oauth-server\\nOAuthServerServiceEndpointAccessibleControllerDegraded: Get \\\"https://172.30.202.245:443/healthz\\\": dial tcp 172.30.202.245:443: connect: connection refused\\nRouterCertsDegraded: neither the custom secret/v4-0-config-system-router-certs -n openshift-authentication or default secret/v4-0-config-system-custom-router-certs -n openshift-authentication could be retrieved: secret \\\"v4-0-config-system-router-certs\\\" not found\\nOpenshiftAuthenticationStaticResourcesDegraded: \\\"oauth-openshift/audit-policy.yaml\\\" (string): namespaces \\\"openshift-authentication\\\" not found\\nOpenshiftAuthenticationStaticResourcesDegraded: \" to \"OAuthServerServiceEndpointsEndpointAccessibleControllerDegraded: oauth service endpoints are not ready\\nWebhookAuthenticatorControllerDegraded: failed to read service-ca crt bundle: open /var/run/configmaps/service-ca-bundle/service-ca.crt: no such file or directory\\nOAuthServerConfigObservationDegraded: secret \\\"v4-0-config-system-router-certs\\\" not found\\nIngressStateEndpointsDegraded: No endpoints found for oauth-server\\nOAuthServerServiceEndpointAccessibleControllerDegraded: Get \\\"https://172.30.202.245:443/healthz\\\": dial tcp 172.30.202.245:443: connect: connection refused\\nRouterCertsDegraded: neither the custom secret/v4-0-config-system-router-certs -n openshift-authentication or default secret/v4-0-config-system-custom-router-certs -n openshift-authentication could be retrieved: secret \\\"v4-0-config-system-router-certs\\\" not found\\nOpenshiftAuthenticationStaticResourcesDegraded: \\\"oauth-openshift/audit-policy.yaml\\\" (string): namespaces \\\"openshift-authentication\\\" not found\\nOpenshiftAuthenticationStaticResourcesDegraded: \"","type":"Normal"},{"namespace":"openshift-authentication-operator","lastTimestamp":"2025-12-04T11:38:19Z","reason":"OperatorStatusChanged","message":"Status for clusteroperator/authentication changed: Degraded message changed from \"OAuthServerServiceEndpointsEndpointAccessibleControllerDegraded: oauth service endpoints are not ready\\nWebhookAuthenticatorControllerDegraded: failed to read service-ca crt bundle: open /var/run/configmaps/service-ca-bundle/service-ca.crt: no such file or directory\\nOAuthServerConfigObservationDegraded: secret \\\"v4-0-config-system-router-certs\\\" not found\\nIngressStateEndpointsDegraded: No endpoints found for oauth-server\\nOAuthServerServiceEndpointAccessibleControllerDegraded: Get \\\"https://172.30.202.245:443/healthz\\\": dial tcp 172.30.202.245:443: connect: connection refused\\nRouterCertsDegraded: neither the custom secret/v4-0-config-system-router-certs -n openshift-authentication or default secret/v4-0-config-system-custom-router-certs -n openshift-authentication could be retrieved: secret \\\"v4-0-config-system-router-certs\\\" not found\\nOpenshiftAuthenticationStaticResourcesDegraded: \\\"oauth-openshift/audit-policy.yaml\\\" (string): namespaces \\\"openshift-authentication\\\" not found\\nOpenshiftAuthenticationStaticResourcesDegraded: \" to \"OAuthServerServiceEndpointsEndpointAccessibleControllerDegraded: oauth service endpoints are not ready\\nWebhookAuthenticatorControllerDegraded: failed to read service-ca crt bundle: open /var/run/configmaps/service-ca-bundle/service-ca.crt: no such file or directory\\nOAuthServerConfigObservationDegraded: secret \\\"v4-0-config-system-router-certs\\\" not found\\nIngressStateEndpointsDegraded: No subsets found for the endpoints of oauth-server\\nOAuthServerServiceEndpointAccessibleControllerDegraded: Get \\\"https://172.30.202.245:443/healthz\\\": dial tcp 172.30.202.245:443: connect: connection refused\\nRouterCertsDegraded: neither the custom secret/v4-0-config-system-router-certs -n openshift-authentication or default secret/v4-0-config-system-custom-router-certs -n openshift-authentication could be retrieved: secret \\\"v4-0-config-system-router-certs\\\" not found\\nOpenshiftAuthenticationStaticResourcesDegraded: \\\"oauth-openshift/audit-policy.yaml\\\" (string): namespaces \\\"openshift-authentication\\\" not found\\nOpenshiftAuthenticationStaticResourcesDegraded: \"","type":"Normal"},{"namespace":"openshift-authentication-operator","lastTimestamp":"2025-12-04T11:38:19Z","reason":"OperatorStatusChanged","message":"Status for clusteroperator/authentication changed: Progressing message changed from \"APIServerDeploymentProgressing: deployment/apiserver.openshift-oauth-apiserver: observed generation is 0, desired generation is 1.\" to \"APIServerDeploymentProgressing: deployment/apiserver.openshift-oauth-apiserver: 1/1 pods have been updated to the latest generation and 0/1 pods are available\"","type":"Normal"},{"namespace":"openshift-authentication-operator","lastTimestamp":"2025-12-04T11:38:20Z","reason":"ConfigMapCreated","message":"Created ConfigMap/audit -n openshift-authentication because it was missing","type":"Normal"},{"namespace":"openshift-authentication-operator","lastTimestamp":"2025-12-04T11:38:24Z","reason":"OperatorStatusChanged","message":"Status for clusteroperator/authentication changed: Degraded message changed from \"OAuthServerServiceEndpointsEndpointAccessibleControllerDegraded: oauth service endpoints are not ready\\nWebhookAuthenticatorControllerDegraded: failed to read service-ca crt bundle: open /var/run/configmaps/service-ca-bundle/service-ca.crt: no such file or directory\\nOAuthServerConfigObservationDegraded: secret \\\"v4-0-config-system-router-certs\\\" not found\\nIngressStateEndpointsDegraded: No subsets found for the endpoints of oauth-server\\nOAuthServerServiceEndpointAccessibleControllerDegraded: Get \\\"https://172.30.202.245:443/healthz\\\": dial tcp 172.30.202.245:443: connect: connection refused\\nRouterCertsDegraded: neither the custom secret/v4-0-config-system-router-certs -n openshift-authentication or default secret/v4-0-config-system-custom-router-certs -n openshift-authentication could be retrieved: secret \\\"v4-0-config-system-router-certs\\\" not found\\nOpenshiftAuthenticationStaticResourcesDegraded: \\\"oauth-openshift/audit-policy.yaml\\\" (string): namespaces \\\"openshift-authentication\\\" not found\\nOpenshiftAuthenticationStaticResourcesDegraded: \" to \"OAuthServerServiceEndpointsEndpointAccessibleControllerDegraded: oauth service endpoints are not ready\\nWebhookAuthenticatorControllerDegraded: failed to read service-ca crt bundle: open /var/run/configmaps/service-ca-bundle/service-ca.crt: no such file or directory\\nOAuthServerConfigObservationDegraded: secret \\\"v4-0-config-system-router-certs\\\" not found\\nIngressStateEndpointsDegraded: No subsets found for the endpoints of oauth-server\\nOAuthServerServiceEndpointAccessibleControllerDegraded: Get \\\"https://172.30.202.245:443/healthz\\\": dial tcp 172.30.202.245:443: connect: connection refused\\nRouterCertsDegraded: neither the custom secret/v4-0-config-system-router-certs -n openshift-authentication or default secret/v4-0-config-system-custom-router-certs -n openshift-authentication could be retrieved: secret \\\"v4-0-config-system-router-certs\\\" not found\"","type":"Normal"},{"namespace":"openshift-authentication-operator","lastTimestamp":"2025-12-04T11:38:25Z","reason":"SecretCreated","message":"Created Secret/v4-0-config-system-router-certs -n openshift-authentication because it was missing","type":"Normal"},{"namespace":"openshift-authentication-operator","lastTimestamp":"2025-12-04T11:38:25Z","reason":"ObserveRouterSecret","message":"namedCertificates changed to []interface {}{map[string]interface {}{\"certFile\":\"/var/config/system/secrets/v4-0-config-system-router-certs/apps.sno.openstack.lab\", \"keyFile\":\"/var/config/system/secrets/v4-0-config-system-router-certs/apps.sno.openstack.lab\", \"names\":[]interface {}{\"*.apps.sno.openstack.lab\"}}}","type":"Normal"},{"namespace":"openshift-authentication-operator","lastTimestamp":"2025-12-04T11:38:25Z","reason":"ObservedConfigChanged","message":"Writing updated section (\"oauthServer\") of observed config: \" map[string]any{\\n \\t\\\"corsAllowedOrigins\\\": []any{string(`//127\\\\.0\\\\.0\\\\.1(:|$)`), string(\\\"//localhost(:|$)\\\")},\\n \\t\\\"oauthConfig\\\": map[string]any{\\\"assetPublicURL\\\": string(\\\"\\\"), \\\"loginURL\\\": string(\\\"https://api.sno.openstack.lab:6443\\\"), \\\"templates\\\": map[string]any{\\\"error\\\": string(\\\"/var/config/system/secrets/v4-0-config-system-ocp-branding-templ\\\"...), \\\"login\\\": string(\\\"/var/config/system/secrets/v4-0-config-system-ocp-branding-templ\\\"...), \\\"providerSelection\\\": string(\\\"/var/config/system/secrets/v4-0-config-system-ocp-branding-templ\\\"...)}, \\\"tokenConfig\\\": map[string]any{\\\"accessTokenMaxAgeSeconds\\\": float64(86400), \\\"authorizeTokenMaxAgeSeconds\\\": float64(300)}},\\n \\t\\\"serverArguments\\\": map[string]any{\\\"audit-log-format\\\": []any{string(\\\"json\\\")}, \\\"audit-log-maxbackup\\\": []any{string(\\\"10\\\")}, \\\"audit-log-maxsize\\\": []any{string(\\\"100\\\")}, \\\"audit-log-path\\\": []any{string(\\\"/var/log/oauth-server/audit.log\\\")}, ...},\\n \\t\\\"servingInfo\\\": map[string]any{\\n \\t\\t\\\"cipherSuites\\\": []any{string(\\\"TLS_AES_128_GCM_SHA256\\\"), string(\\\"TLS_AES_256_GCM_SHA384\\\"), string(\\\"TLS_CHACHA20_POLY1305_SHA256\\\"), string(\\\"TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256\\\"), ...},\\n \\t\\t\\\"minTLSVersion\\\": string(\\\"VersionTLS12\\\"),\\n+ \\t\\t\\\"namedCertificates\\\": []any{\\n+ \\t\\t\\tmap[string]any{\\n+ \\t\\t\\t\\t\\\"certFile\\\": string(\\\"/var/config/system/secrets/v4-0-config-system-router-certs/apps.\\\"...),\\n+ \\t\\t\\t\\t\\\"keyFile\\\": string(\\\"/var/config/system/secrets/v4-0-config-system-router-certs/apps.\\\"...),\\n+ \\t\\t\\t\\t\\\"names\\\": []any{string(\\\"*.apps.sno.openstack.lab\\\")},\\n+ \\t\\t\\t},\\n+ \\t\\t},\\n \\t},\\n \\t\\\"volumesToMount\\\": map[string]any{\\\"identityProviders\\\": string(\\\"{}\\\")},\\n }\\n\"","type":"Normal"},{"namespace":"openshift-authentication-operator","lastTimestamp":"2025-12-04T11:38:25Z","reason":"OperatorStatusChanged","message":"Status for clusteroperator/authentication changed: Degraded message changed from \"OAuthServerServiceEndpointsEndpointAccessibleControllerDegraded: oauth service endpoints are not ready\\nWebhookAuthenticatorControllerDegraded: failed to read service-ca crt bundle: open /var/run/configmaps/service-ca-bundle/service-ca.crt: no such file or directory\\nOAuthServerConfigObservationDegraded: secret \\\"v4-0-config-system-router-certs\\\" not found\\nIngressStateEndpointsDegraded: No subsets found for the endpoints of oauth-server\\nOAuthServerServiceEndpointAccessibleControllerDegraded: Get \\\"https://172.30.202.245:443/healthz\\\": dial tcp 172.30.202.245:443: connect: connection refused\\nRouterCertsDegraded: neither the custom secret/v4-0-config-system-router-certs -n openshift-authentication or default secret/v4-0-config-system-custom-router-certs -n openshift-authentication could be retrieved: secret \\\"v4-0-config-system-router-certs\\\" not found\" to \"OAuthServerServiceEndpointsEndpointAccessibleControllerDegraded: oauth service endpoints are not ready\\nWebhookAuthenticatorControllerDegraded: failed to read service-ca crt bundle: open /var/run/configmaps/service-ca-bundle/service-ca.crt: no such file or directory\\nOAuthServerConfigObservationDegraded: secret \\\"v4-0-config-system-router-certs\\\" not found\\nIngressStateEndpointsDegraded: No subsets found for the endpoints of oauth-server\\nOAuthServerServiceEndpointAccessibleControllerDegraded: Get \\\"https://172.30.202.245:443/healthz\\\": dial tcp 172.30.202.245:443: connect: connection refused\"","type":"Normal"},{"namespace":"openshift-authentication-operator","lastTimestamp":"2025-12-04T11:38:25Z","reason":"OperatorStatusChanged","message":"Status for clusteroperator/authentication changed: Degraded message changed from \"OAuthServerServiceEndpointsEndpointAccessibleControllerDegraded: oauth service endpoints are not ready\\nWebhookAuthenticatorControllerDegraded: failed to read service-ca crt bundle: open /var/run/configmaps/service-ca-bundle/service-ca.crt: no such file or directory\\nOAuthServerConfigObservationDegraded: secret \\\"v4-0-config-system-router-certs\\\" not found\\nIngressStateEndpointsDegraded: No subsets found for the endpoints of oauth-server\\nOAuthServerServiceEndpointAccessibleControllerDegraded: Get \\\"https://172.30.202.245:443/healthz\\\": dial tcp 172.30.202.245:443: connect: connection refused\" to \"OAuthServerServiceEndpointsEndpointAccessibleControllerDegraded: oauth service endpoints are not ready\\nWebhookAuthenticatorControllerDegraded: failed to read service-ca crt bundle: open /var/run/configmaps/service-ca-bundle/service-ca.crt: no such file or directory\\nIngressStateEndpointsDegraded: No subsets found for the endpoints of oauth-server\\nOAuthServerServiceEndpointAccessibleControllerDegraded: Get \\\"https://172.30.202.245:443/healthz\\\": dial tcp 172.30.202.245:443: connect: connection refused\"","type":"Normal"},{"namespace":"openshift-authentication-operator","lastTimestamp":"2025-12-04T11:38:28Z","reason":"ConfigMapCreated","message":"Created ConfigMap/oauth-serving-cert -n openshift-config-managed because it was missing","type":"Normal"},{"namespace":"openshift-authentication-operator","lastTimestamp":"2025-12-04T11:38:41Z","reason":"\u003cunknown\u003eCreated","message":"Created \u003cunknown\u003e/v1.oauth.openshift.io because it was missing","type":"Normal"},{"namespace":"openshift-authentication-operator","lastTimestamp":"2025-12-04T11:38:41Z","reason":"\u003cunknown\u003eCreated","message":"Created \u003cunknown\u003e/v1.user.openshift.io because it was missing","type":"Normal"},{"namespace":"openshift-authentication-operator","lastTimestamp":"2025-12-04T11:38:41Z","reason":"OpenShiftAPICheckFailed","message":"\"oauth.openshift.io.v1\" failed with an attempt failed with statusCode = 503, err = the server is currently unable to handle the request","type":"Warning"},{"namespace":"openshift-authentication-operator","lastTimestamp":"2025-12-04T11:38:41Z","reason":"OpenShiftAPICheckFailed","message":"\"user.openshift.io.v1\" failed with an attempt failed with statusCode = 503, err = the server is currently unable to handle the request","type":"Warning"},{"namespace":"openshift-authentication-operator","lastTimestamp":"2025-12-04T11:38:41Z","reason":"OperatorStatusChanged","message":"Status for clusteroperator/authentication changed: Available message changed from \"APIServerDeploymentAvailable: no apiserver.openshift-oauth-apiserver pods available on any node.\\nAPIServicesAvailable: PreconditionNotReady\\nOAuthServerServiceEndpointAccessibleControllerAvailable: Get \\\"https://172.30.202.245:443/healthz\\\": dial tcp 172.30.202.245:443: connect: connection refused\\nOAuthServerServiceEndpointsEndpointAccessibleControllerAvailable: endpoints \\\"oauth-openshift\\\" not found\\nReadyIngressNodesAvailable: Authentication requires functional ingress which requires at least one schedulable and ready node. Got 0 worker nodes, 1 master nodes, 0 custom target nodes (none are schedulable or ready for ingress pods).\" to \"APIServerDeploymentAvailable: no apiserver.openshift-oauth-apiserver pods available on any node.\\nAPIServicesAvailable: \\\"oauth.openshift.io.v1\\\" is not ready: an attempt failed with statusCode = 503, err = the server is currently unable to handle the request\\nAPIServicesAvailable: \\\"user.openshift.io.v1\\\" is not ready: an attempt failed with statusCode = 503, err = the server is currently unable to handle the request\\nOAuthServerServiceEndpointAccessibleControllerAvailable: Get \\\"https://172.30.202.245:443/healthz\\\": dial tcp 172.30.202.245:443: connect: connection refused\\nOAuthServerServiceEndpointsEndpointAccessibleControllerAvailable: endpoints \\\"oauth-openshift\\\" not found\\nReadyIngressNodesAvailable: Authentication requires functional ingress which requires at least one schedulable and ready node. Got 0 worker nodes, 1 master nodes, 0 custom target nodes (none are schedulable or ready for ingress pods).\"","type":"Normal"},{"namespace":"openshift-authentication-operator","lastTimestamp":"2025-12-04T11:38:41Z","reason":"OperatorStatusChanged","message":"Status for clusteroperator/authentication changed: Available message changed from \"APIServerDeploymentAvailable: no apiserver.openshift-oauth-apiserver pods available on any node.\\nAPIServicesAvailable: \\\"oauth.openshift.io.v1\\\" is not ready: an attempt failed with statusCode = 503, err = the server is currently unable to handle the request\\nAPIServicesAvailable: \\\"user.openshift.io.v1\\\" is not ready: an attempt failed with statusCode = 503, err = the server is currently unable to handle the request\\nOAuthServerServiceEndpointAccessibleControllerAvailable: Get \\\"https://172.30.202.245:443/healthz\\\": dial tcp 172.30.202.245:443: connect: connection refused\\nOAuthServerServiceEndpointsEndpointAccessibleControllerAvailable: endpoints \\\"oauth-openshift\\\" not found\\nReadyIngressNodesAvailable: Authentication requires functional ingress which requires at least one schedulable and ready node. Got 0 worker nodes, 1 master nodes, 0 custom target nodes (none are schedulable or ready for ingress pods).\" to \"APIServerDeploymentAvailable: no apiserver.openshift-oauth-apiserver pods available on any node.\\nAPIServicesAvailable: apiservices.apiregistration.k8s.io/v1.oauth.openshift.io: not available: failing or missing response from https://10.128.0.39:8443/apis/oauth.openshift.io/v1: bad status from https://10.128.0.39:8443/apis/oauth.openshift.io/v1: 401\\nAPIServicesAvailable: apiservices.apiregistration.k8s.io/v1.user.openshift.io: not available: failing or missing response from https://10.128.0.39:8443/apis/user.openshift.io/v1: bad status from https://10.128.0.39:8443/apis/user.openshift.io/v1: 401\\nOAuthServerServiceEndpointAccessibleControllerAvailable: Get \\\"https://172.30.202.245:443/healthz\\\": dial tcp 172.30.202.245:443: connect: connection refused\\nOAuthServerServiceEndpointsEndpointAccessibleControllerAvailable: endpoints \\\"oauth-openshift\\\" not found\\nReadyIngressNodesAvailable: Authentication requires functional ingress which requires at least one schedulable and ready node. Got 0 worker nodes, 1 master nodes, 0 custom target nodes (none are schedulable or ready for ingress pods).\"","type":"Normal"},{"namespace":"openshift-authentication-operator","lastTimestamp":"2025-12-04T11:38:43Z","reason":"OperatorVersionChanged","message":"clusteroperator/authentication version \"oauth-apiserver\" changed from \"\" to \"4.18.29\"","type":"Normal"},{"namespace":"openshift-authentication-operator","lastTimestamp":"2025-12-04T11:38:43Z","reason":"OperatorStatusChanged","message":"Status for clusteroperator/authentication changed: status.versions changed from [{\"operator\" \"4.18.29\"}] to [{\"operator\" \"4.18.29\"} {\"oauth-apiserver\" \"4.18.29\"}]","type":"Normal"},{"namespace":"openshift-authentication-operator","lastTimestamp":"2025-12-04T11:38:43Z","reason":"OperatorStatusChanged","message":"Status for clusteroperator/authentication changed: Progressing changed from True to False (\"AuthenticatorCertKeyProgressing: All is well\"),Available message changed from \"APIServerDeploymentAvailable: no apiserver.openshift-oauth-apiserver pods available on any node.\\nAPIServicesAvailable: apiservices.apiregistration.k8s.io/v1.oauth.openshift.io: not available: failing or missing response from https://10.128.0.39:8443/apis/oauth.openshift.io/v1: bad status from https://10.128.0.39:8443/apis/oauth.openshift.io/v1: 401\\nAPIServicesAvailable: apiservices.apiregistration.k8s.io/v1.user.openshift.io: not available: failing or missing response from https://10.128.0.39:8443/apis/user.openshift.io/v1: bad status from https://10.128.0.39:8443/apis/user.openshift.io/v1: 401\\nOAuthServerServiceEndpointAccessibleControllerAvailable: Get \\\"https://172.30.202.245:443/healthz\\\": dial tcp 172.30.202.245:443: connect: connection refused\\nOAuthServerServiceEndpointsEndpointAccessibleControllerAvailable: endpoints \\\"oauth-openshift\\\" not found\\nReadyIngressNodesAvailable: Authentication requires functional ingress which requires at least one schedulable and ready node. Got 0 worker nodes, 1 master nodes, 0 custom target nodes (none are schedulable or ready for ingress pods).\" to \"APIServicesAvailable: apiservices.apiregistration.k8s.io/v1.oauth.openshift.io: not available: failing or missing response from https://10.128.0.39:8443/apis/oauth.openshift.io/v1: bad status from https://10.128.0.39:8443/apis/oauth.openshift.io/v1: 401\\nAPIServicesAvailable: apiservices.apiregistration.k8s.io/v1.user.openshift.io: not available: failing or missing response from https://10.128.0.39:8443/apis/user.openshift.io/v1: bad status from https://10.128.0.39:8443/apis/user.openshift.io/v1: 401\\nOAuthServerServiceEndpointAccessibleControllerAvailable: Get \\\"https://172.30.202.245:443/healthz\\\": dial tcp 172.30.202.245:443: connect: connection refused\\nOAuthServerServiceEndpointsEndpointAccessibleControllerAvailable: endpoints \\\"oauth-openshift\\\" not found\\nReadyIngressNodesAvailable: Authentication requires functional ingress which requires at least one schedulable and ready node. Got 0 worker nodes, 1 master nodes, 0 custom target nodes (none are schedulable or ready for ingress pods).\"","type":"Normal"},{"namespace":"openshift-authentication-operator","lastTimestamp":"2025-12-04T11:43:09Z","reason":"Unhealthy","message":"Liveness probe failed: Get \"https://10.128.0.11:8443/healthz\": dial tcp 10.128.0.11:8443: connect: connection refused","type":"Warning"},{"namespace":"openshift-authentication-operator","lastTimestamp":"2025-12-04T11:43:09Z","reason":"ProbeError","message":"Liveness probe error: Get \"https://10.128.0.11:8443/healthz\": dial tcp 10.128.0.11:8443: connect: connection refused\nbody: \n","type":"Warning"},{"namespace":"openshift-authentication-operator","lastTimestamp":"2025-12-04T11:43:53Z","reason":"BackOff","message":"Back-off restarting failed container authentication-operator in pod authentication-operator-6c968fdfdf-nrrfw_openshift-authentication-operator(42b3be0f-1d82-4a64-abb4-0118a6960efd)","type":"Warning"},{"namespace":"openshift-authentication-operator","lastTimestamp":"2025-12-04T11:44:27Z","reason":"Pulled","message":"Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e85850a4ae1a1e3ec2c590a4936d640882b6550124da22031c85b526afbf52df\" already present on machine","type":"Normal"},{"namespace":"openshift-authentication-operator","lastTimestamp":"2025-12-04T11:44:28Z","reason":"Created","message":"Created container: authentication-operator","type":"Normal"},{"namespace":"openshift-authentication-operator","lastTimestamp":"2025-12-04T11:44:28Z","reason":"Started","message":"Started container authentication-operator","type":"Normal"},{"namespace":"openshift-authentication-operator","lastTimestamp":"2025-12-04T11:56:29Z","reason":"LeaderElection","message":"authentication-operator-6c968fdfdf-nrrfw_ed095787-ae9c-47b2-8f37-9613c2bc84fa became leader","type":"Normal"},{"namespace":"openshift-authentication-operator","lastTimestamp":"2025-12-04T11:56:30Z","reason":"OperatorStatusChanged","message":"Status for clusteroperator/authentication changed: Degraded changed from False to True (\"IngressStateEndpointsDegraded: No subsets found for the endpoints of oauth-server\\nOAuthServerServiceEndpointAccessibleControllerDegraded: Get \\\"https://172.30.202.245:443/healthz\\\": dial tcp 172.30.202.245:443: connect: connection refused\\nOAuthServerServiceEndpointsEndpointAccessibleControllerDegraded: oauth service endpoints are not ready\\nWebhookAuthenticatorControllerDegraded: failed to read service-ca crt bundle: open /var/run/configmaps/service-ca-bundle/service-ca.crt: no such file or directory\")","type":"Normal"},{"namespace":"openshift-authentication-operator","lastTimestamp":"2025-12-04T11:56:34Z","reason":"FastControllerResync","message":"Controller \"openshift-apiserver-APIService\" resync interval is set to 10s which might lead to client request throttling","type":"Warning"},{"namespace":"openshift-authentication-operator","lastTimestamp":"2025-12-04T11:56:34Z","reason":"FastControllerResync","message":"Controller \"auditPolicyController\" resync interval is set to 10s which might lead to client request throttling","type":"Warning"},{"namespace":"openshift-authentication-operator","lastTimestamp":"2025-12-04T11:56:38Z","reason":"SecretCreated","message":"Created Secret/webhook-authentication-integrated-oauth -n openshift-config because it was missing","type":"Normal"},{"namespace":"openshift-authentication-operator","lastTimestamp":"2025-12-04T11:56:39Z","reason":"OperatorStatusChanged","message":"Status for clusteroperator/authentication changed: Degraded message changed from \"IngressStateEndpointsDegraded: No subsets found for the endpoints of oauth-server\\nOAuthServerServiceEndpointAccessibleControllerDegraded: Get \\\"https://172.30.202.245:443/healthz\\\": dial tcp 172.30.202.245:443: connect: connection refused\\nOAuthServerServiceEndpointsEndpointAccessibleControllerDegraded: oauth service endpoints are not ready\\nWebhookAuthenticatorControllerDegraded: failed to read service-ca crt bundle: open /var/run/configmaps/service-ca-bundle/service-ca.crt: no such file or directory\" to \"IngressStateEndpointsDegraded: No subsets found for the endpoints of oauth-server\\nOAuthServerServiceEndpointAccessibleControllerDegraded: Get \\\"https://172.30.202.245:443/healthz\\\": dial tcp 172.30.202.245:443: connect: connection refused\\nOAuthServerServiceEndpointsEndpointAccessibleControllerDegraded: oauth service endpoints are not ready\"","type":"Normal"},{"namespace":"openshift-authentication-operator","lastTimestamp":"2025-12-04T11:59:49Z","reason":"SecretCreated","message":"Created Secret/v4-0-config-system-session -n openshift-authentication because it was missing","type":"Normal"},{"namespace":"openshift-authentication-operator","lastTimestamp":"2025-12-04T11:59:49Z","reason":"ConfigMapCreated","message":"Created ConfigMap/v4-0-config-system-cliconfig -n openshift-authentication because it was missing","type":"Normal"},{"namespace":"openshift-authentication-operator","lastTimestamp":"2025-12-04T11:59:50Z","reason":"OperatorStatusChanged","message":"Status for clusteroperator/authentication changed: Degraded message changed from \"IngressStateEndpointsDegraded: No subsets found for the endpoints of oauth-server\\nOAuthServerServiceEndpointAccessibleControllerDegraded: Get \\\"https://172.30.202.245:443/healthz\\\": dial tcp 172.30.202.245:443: connect: connection refused\\nOAuthServerServiceEndpointsEndpointAccessibleControllerDegraded: oauth service endpoints are not ready\" to \"IngressStateEndpointsDegraded: No subsets found for the endpoints of oauth-server\\nOAuthServerDeploymentDegraded: waiting for the oauth-openshift route to contain an admitted ingress: no admitted ingress for route oauth-openshift in namespace openshift-authentication\\nOAuthServerDeploymentDegraded: \\nOAuthServerRouteEndpointAccessibleControllerDegraded: ingress.config/cluster does not yet have status for the \\\"openshift-authentication/oauth-openshift\\\" route\\nOAuthServerServiceEndpointAccessibleControllerDegraded: Get \\\"https://172.30.202.245:443/healthz\\\": dial tcp 172.30.202.245:443: connect: connection refused\\nOAuthServerServiceEndpointsEndpointAccessibleControllerDegraded: oauth service endpoints are not ready\\nOAuthServerWorkloadDegraded: waiting for the oauth-openshift route to contain an admitted ingress: no admitted ingress for route oauth-openshift in namespace openshift-authentication\\nOAuthServerWorkloadDegraded: \\nWellKnownReadyControllerDegraded: route.route.openshift.io \\\"oauth-openshift\\\" not found\",Progressing changed from False to True (\"\"),Available message changed from \"APIServicesAvailable: apiservices.apiregistration.k8s.io/v1.oauth.openshift.io: not available: failing or missing response from https://10.128.0.39:8443/apis/oauth.openshift.io/v1: bad status from https://10.128.0.39:8443/apis/oauth.openshift.io/v1: 401\\nAPIServicesAvailable: apiservices.apiregistration.k8s.io/v1.user.openshift.io: not available: failing or missing response from https://10.128.0.39:8443/apis/user.openshift.io/v1: bad status from https://10.128.0.39:8443/apis/user.openshift.io/v1: 401\\nOAuthServerServiceEndpointAccessibleControllerAvailable: Get \\\"https://172.30.202.245:443/healthz\\\": dial tcp 172.30.202.245:443: connect: connection refused\\nOAuthServerServiceEndpointsEndpointAccessibleControllerAvailable: endpoints \\\"oauth-openshift\\\" not found\\nReadyIngressNodesAvailable: Authentication requires functional ingress which requires at least one schedulable and ready node. Got 0 worker nodes, 1 master nodes, 0 custom target nodes (none are schedulable or ready for ingress pods).\" to \"OAuthServerRouteEndpointAccessibleControllerAvailable: failed to retrieve route from cache: route.route.openshift.io \\\"oauth-openshift\\\" not found\\nOAuthServerServiceEndpointAccessibleControllerAvailable: Get \\\"https://172.30.202.245:443/healthz\\\": dial tcp 172.30.202.245:443: connect: connection refused\\nOAuthServerServiceEndpointsEndpointAccessibleControllerAvailable: endpoints \\\"oauth-openshift\\\" not found\\nReadyIngressNodesAvailable: Authentication requires functional ingress which requires at least one schedulable and ready node. Got 0 worker nodes, 1 master nodes, 0 custom target nodes (none are schedulable or ready for ingress pods).\\nWellKnownAvailable: route.route.openshift.io \\\"oauth-openshift\\\" not found\"","type":"Normal"},{"namespace":"openshift-authentication-operator","lastTimestamp":"2025-12-04T11:59:57Z","reason":"OperatorStatusChanged","message":"Status for clusteroperator/authentication changed: Degraded message changed from \"IngressStateEndpointsDegraded: No subsets found for the endpoints of oauth-server\\nOAuthServerDeploymentDegraded: waiting for the oauth-openshift route to contain an admitted ingress: no admitted ingress for route oauth-openshift in namespace openshift-authentication\\nOAuthServerDeploymentDegraded: \\nOAuthServerRouteEndpointAccessibleControllerDegraded: ingress.config/cluster does not yet have status for the \\\"openshift-authentication/oauth-openshift\\\" route\\nOAuthServerServiceEndpointAccessibleControllerDegraded: Get \\\"https://172.30.202.245:443/healthz\\\": dial tcp 172.30.202.245:443: connect: connection refused\\nOAuthServerServiceEndpointsEndpointAccessibleControllerDegraded: oauth service endpoints are not ready\\nOAuthServerWorkloadDegraded: waiting for the oauth-openshift route to contain an admitted ingress: no admitted ingress for route oauth-openshift in namespace openshift-authentication\\nOAuthServerWorkloadDegraded: \\nWellKnownReadyControllerDegraded: route.route.openshift.io \\\"oauth-openshift\\\" not found\" to \"IngressStateEndpointsDegraded: No subsets found for the endpoints of oauth-server\\nOAuthServerDeploymentDegraded: waiting for the oauth-openshift route to contain an admitted ingress: no admitted ingress for route oauth-openshift in namespace openshift-authentication\\nOAuthServerDeploymentDegraded: \\nOAuthServerRouteEndpointAccessibleControllerDegraded: ingress.config/cluster does not yet have status for the \\\"openshift-authentication/oauth-openshift\\\" route\\nOAuthServerServiceEndpointAccessibleControllerDegraded: Get \\\"https://172.30.202.245:443/healthz\\\": dial tcp 172.30.202.245:443: connect: connection refused\\nOAuthServerServiceEndpointsEndpointAccessibleControllerDegraded: oauth service endpoints are not ready\\nOAuthServerWorkloadDegraded: waiting for the oauth-openshift route to contain an admitted ingress: no admitted ingress for route oauth-openshift in namespace openshift-authentication\\nOAuthServerWorkloadDegraded: \\nWellKnownReadyControllerDegraded: failed to get oauth metadata from openshift-config-managed/oauth-openshift ConfigMap: configmap \\\"oauth-openshift\\\" not found (check authentication operator, it is supposed to create this)\",Available message changed from \"OAuthServerRouteEndpointAccessibleControllerAvailable: failed to retrieve route from cache: route.route.openshift.io \\\"oauth-openshift\\\" not found\\nOAuthServerServiceEndpointAccessibleControllerAvailable: Get \\\"https://172.30.202.245:443/healthz\\\": dial tcp 172.30.202.245:443: connect: connection refused\\nOAuthServerServiceEndpointsEndpointAccessibleControllerAvailable: endpoints \\\"oauth-openshift\\\" not found\\nReadyIngressNodesAvailable: Authentication requires functional ingress which requires at least one schedulable and ready node. Got 0 worker nodes, 1 master nodes, 0 custom target nodes (none are schedulable or ready for ingress pods).\\nWellKnownAvailable: route.route.openshift.io \\\"oauth-openshift\\\" not found\" to \"OAuthServerRouteEndpointAccessibleControllerAvailable: failed to retrieve route from cache: route.route.openshift.io \\\"oauth-openshift\\\" not found\\nOAuthServerServiceEndpointAccessibleControllerAvailable: Get \\\"https://172.30.202.245:443/healthz\\\": dial tcp 172.30.202.245:443: connect: connection refused\\nOAuthServerServiceEndpointsEndpointAccessibleControllerAvailable: endpoints \\\"oauth-openshift\\\" not found\\nReadyIngressNodesAvailable: Authentication requires functional ingress which requires at least one schedulable and ready node. Got 0 worker nodes, 1 master nodes, 0 custom target nodes (none are schedulable or ready for ingress pods).\\nWellKnownAvailable: The well-known endpoint is not yet available: failed to get oauth metadata from openshift-config-managed/oauth-openshift ConfigMap: configmap \\\"oauth-openshift\\\" not found (check authentication operator, it is supposed to create this)\"","type":"Normal"},{"namespace":"openshift-authentication-operator","lastTimestamp":"2025-12-04T11:59:58Z","reason":"OperatorStatusChanged","message":"Status for clusteroperator/authentication changed: Degraded message changed from \"IngressStateEndpointsDegraded: No subsets found for the endpoints of oauth-server\\nOAuthServerDeploymentDegraded: waiting for the oauth-openshift route to contain an admitted ingress: no admitted ingress for route oauth-openshift in namespace openshift-authentication\\nOAuthServerDeploymentDegraded: \\nOAuthServerRouteEndpointAccessibleControllerDegraded: ingress.config/cluster does not yet have status for the \\\"openshift-authentication/oauth-openshift\\\" route\\nOAuthServerServiceEndpointAccessibleControllerDegraded: Get \\\"https://172.30.202.245:443/healthz\\\": dial tcp 172.30.202.245:443: connect: connection refused\\nOAuthServerServiceEndpointsEndpointAccessibleControllerDegraded: oauth service endpoints are not ready\\nOAuthServerWorkloadDegraded: waiting for the oauth-openshift route to contain an admitted ingress: no admitted ingress for route oauth-openshift in namespace openshift-authentication\\nOAuthServerWorkloadDegraded: \\nWellKnownReadyControllerDegraded: failed to get oauth metadata from openshift-config-managed/oauth-openshift ConfigMap: configmap \\\"oauth-openshift\\\" not found (check authentication operator, it is supposed to create this)\" to \"IngressStateEndpointsDegraded: No subsets found for the endpoints of oauth-server\\nOAuthServerDeploymentDegraded: waiting for the oauth-openshift route to contain an admitted ingress: no admitted ingress for route oauth-openshift in namespace openshift-authentication\\nOAuthServerDeploymentDegraded: \\nOAuthServerRouteEndpointAccessibleControllerDegraded: route \\\"openshift-authentication/oauth-openshift\\\": status does not have a valid host address\\nOAuthServerServiceEndpointAccessibleControllerDegraded: Get \\\"https://172.30.202.245:443/healthz\\\": dial tcp 172.30.202.245:443: connect: connection refused\\nOAuthServerServiceEndpointsEndpointAccessibleControllerDegraded: oauth service endpoints are not ready\\nOAuthServerWorkloadDegraded: waiting for the oauth-openshift route to contain an admitted ingress: no admitted ingress for route oauth-openshift in namespace openshift-authentication\\nOAuthServerWorkloadDegraded: \\nWellKnownReadyControllerDegraded: failed to get oauth metadata from openshift-config-managed/oauth-openshift ConfigMap: configmap \\\"oauth-openshift\\\" not found (check authentication operator, it is supposed to create this)\"","type":"Normal"},{"namespace":"openshift-authentication-operator","lastTimestamp":"2025-12-04T12:00:17Z","reason":"OperatorStatusChanged","message":"Status for clusteroperator/authentication changed: Degraded message changed from \"IngressStateEndpointsDegraded: No subsets found for the endpoints of oauth-server\\nOAuthServerDeploymentDegraded: waiting for the oauth-openshift route to contain an admitted ingress: no admitted ingress for route oauth-openshift in namespace openshift-authentication\\nOAuthServerDeploymentDegraded: \\nOAuthServerRouteEndpointAccessibleControllerDegraded: route \\\"openshift-authentication/oauth-openshift\\\": status does not have a valid host address\\nOAuthServerServiceEndpointAccessibleControllerDegraded: Get \\\"https://172.30.202.245:443/healthz\\\": dial tcp 172.30.202.245:443: connect: connection refused\\nOAuthServerServiceEndpointsEndpointAccessibleControllerDegraded: oauth service endpoints are not ready\\nOAuthServerWorkloadDegraded: waiting for the oauth-openshift route to contain an admitted ingress: no admitted ingress for route oauth-openshift in namespace openshift-authentication\\nOAuthServerWorkloadDegraded: \\nWellKnownReadyControllerDegraded: failed to get oauth metadata from openshift-config-managed/oauth-openshift ConfigMap: configmap \\\"oauth-openshift\\\" not found (check authentication operator, it is supposed to create this)\" to \"IngressStateEndpointsDegraded: No subsets found for the endpoints of oauth-server\\nOAuthClientsControllerDegraded: no ingress for host oauth-openshift.apps.sno.openstack.lab in route oauth-openshift in namespace openshift-authentication\\nOAuthServerDeploymentDegraded: waiting for the oauth-openshift route to contain an admitted ingress: no admitted ingress for route oauth-openshift in namespace openshift-authentication\\nOAuthServerDeploymentDegraded: \\nOAuthServerRouteEndpointAccessibleControllerDegraded: route \\\"openshift-authentication/oauth-openshift\\\": status does not have a valid host address\\nOAuthServerServiceEndpointAccessibleControllerDegraded: Get \\\"https://172.30.202.245:443/healthz\\\": dial tcp 172.30.202.245:443: connect: connection refused\\nOAuthServerServiceEndpointsEndpointAccessibleControllerDegraded: oauth service endpoints are not ready\\nOAuthServerWorkloadDegraded: waiting for the oauth-openshift route to contain an admitted ingress: no admitted ingress for route oauth-openshift in namespace openshift-authentication\\nOAuthServerWorkloadDegraded: \\nWellKnownReadyControllerDegraded: failed to get oauth metadata from openshift-config-managed/oauth-openshift ConfigMap: configmap \\\"oauth-openshift\\\" not found (check authentication operator, it is supposed to create this)\"","type":"Normal"}]}events/openshift-config-managed.json0000640000000000000000000000061015114274105016270 0ustar0000000000000000{"items":[{"namespace":"openshift-config-managed","lastTimestamp":"2025-12-04T11:38:22Z","reason":"PublishedRouterCA","message":"Published \"default-ingress-cert\" in \"openshift-config-managed\"","type":"Normal"},{"namespace":"openshift-config-managed","lastTimestamp":"2025-12-04T11:38:22Z","reason":"PublishedRouterCertificates","message":"Published router certificates","type":"Normal"}]}config/serviceaccounts.json0000640000000000000000000003553215114274075014616 0ustar0000000000000000{"serviceAccounts":{"TOTAL_COUNT":346,"namespaces":{"default":[{"name":"builder","secrets":1},{"name":"default","secrets":1},{"name":"deployer","secrets":1}],"kube-public":[{"name":"builder","secrets":1},{"name":"default","secrets":1},{"name":"deployer","secrets":1}],"kube-system":[{"name":"attachdetach-controller","secrets":1},{"name":"builder","secrets":1},{"name":"certificate-controller","secrets":1},{"name":"cloud-controller-manager","secrets":1},{"name":"clusterrole-aggregation-controller","secrets":1},{"name":"cronjob-controller","secrets":1},{"name":"daemon-set-controller","secrets":1},{"name":"default","secrets":1},{"name":"deployer","secrets":1},{"name":"deployment-controller","secrets":1},{"name":"disruption-controller","secrets":1},{"name":"endpoint-controller","secrets":1},{"name":"endpointslice-controller","secrets":1},{"name":"endpointslicemirroring-controller","secrets":1},{"name":"ephemeral-volume-controller","secrets":1},{"name":"expand-controller","secrets":1},{"name":"generic-garbage-collector","secrets":1},{"name":"horizontal-pod-autoscaler","secrets":1},{"name":"job-controller","secrets":1},{"name":"legacy-service-account-token-cleaner","secrets":1},{"name":"namespace-controller","secrets":1},{"name":"node-controller","secrets":1},{"name":"persistent-volume-binder","secrets":1},{"name":"pod-garbage-collector","secrets":1},{"name":"pv-protection-controller","secrets":1},{"name":"pvc-protection-controller","secrets":1},{"name":"replicaset-controller","secrets":1},{"name":"replication-controller","secrets":1},{"name":"resourcequota-controller","secrets":1},{"name":"root-ca-cert-publisher","secrets":1},{"name":"service-account-controller","secrets":1},{"name":"service-ca-cert-publisher","secrets":1},{"name":"statefulset-controller","secrets":1},{"name":"ttl-after-finished-controller","secrets":1},{"name":"validatingadmissionpolicy-status-controller","secrets":1}],"openshift":[{"name":"builder","secrets":1},{"name":"default","secrets":1},{"name":"deployer","secrets":1}],"openshift-apiserver":[{"name":"builder","secrets":1},{"name":"default","secrets":1},{"name":"deployer","secrets":1},{"name":"openshift-apiserver-sa","secrets":1}],"openshift-apiserver-operator":[{"name":"builder","secrets":1},{"name":"default","secrets":1},{"name":"deployer","secrets":1},{"name":"openshift-apiserver-operator","secrets":1}],"openshift-authentication":[{"name":"builder","secrets":1},{"name":"default","secrets":1},{"name":"deployer","secrets":1},{"name":"oauth-openshift","secrets":1}],"openshift-authentication-operator":[{"name":"authentication-operator","secrets":1},{"name":"builder","secrets":1},{"name":"default","secrets":1},{"name":"deployer","secrets":1}],"openshift-catalogd":[{"name":"builder","secrets":1},{"name":"catalogd-controller-manager","secrets":1},{"name":"default","secrets":1},{"name":"deployer","secrets":1}],"openshift-cloud-controller-manager":[{"name":"builder","secrets":1},{"name":"cloud-controller-manager","secrets":1},{"name":"cloud-node-manager","secrets":1},{"name":"default","secrets":1},{"name":"deployer","secrets":1}],"openshift-cloud-controller-manager-operator":[{"name":"builder","secrets":1},{"name":"cluster-cloud-controller-manager","secrets":1},{"name":"default","secrets":1},{"name":"deployer","secrets":1}],"openshift-cloud-credential-operator":[{"name":"builder","secrets":1},{"name":"cloud-credential-operator","secrets":1},{"name":"default","secrets":1},{"name":"deployer","secrets":1}],"openshift-cloud-network-config-controller":[{"name":"builder","secrets":1},{"name":"default","secrets":1},{"name":"deployer","secrets":1}],"openshift-cloud-platform-infra":[{"name":"builder","secrets":1},{"name":"default","secrets":1},{"name":"deployer","secrets":1}],"openshift-cluster-csi-drivers":[{"name":"builder","secrets":1},{"name":"default","secrets":1},{"name":"deployer","secrets":1}],"openshift-cluster-machine-approver":[{"name":"builder","secrets":1},{"name":"default","secrets":1},{"name":"deployer","secrets":1},{"name":"machine-approver-sa","secrets":1}],"openshift-cluster-node-tuning-operator":[{"name":"builder","secrets":1},{"name":"cluster-node-tuning-operator","secrets":1},{"name":"default","secrets":1},{"name":"deployer","secrets":1},{"name":"tuned","secrets":1}],"openshift-cluster-olm-operator":[{"name":"builder","secrets":1},{"name":"cluster-olm-operator","secrets":1},{"name":"default","secrets":1},{"name":"deployer","secrets":1}],"openshift-cluster-samples-operator":[{"name":"builder","secrets":1},{"name":"cluster-samples-operator","secrets":1},{"name":"default","secrets":1},{"name":"deployer","secrets":1}],"openshift-cluster-storage-operator":[{"name":"builder","secrets":1},{"name":"cluster-storage-operator","secrets":1},{"name":"csi-snapshot-controller","secrets":1},{"name":"csi-snapshot-controller-operator","secrets":1},{"name":"default","secrets":1},{"name":"deployer","secrets":1}],"openshift-cluster-version":[{"name":"builder","secrets":1},{"name":"default","secrets":1},{"name":"deployer","secrets":1}],"openshift-config":[{"name":"builder","secrets":1},{"name":"default","secrets":1},{"name":"deployer","secrets":1}],"openshift-config-managed":[{"name":"builder","secrets":1},{"name":"default","secrets":1},{"name":"deployer","secrets":1}],"openshift-config-operator":[{"name":"builder","secrets":1},{"name":"default","secrets":1},{"name":"deployer","secrets":1},{"name":"openshift-config-operator","secrets":1}],"openshift-console":[{"name":"builder","secrets":1},{"name":"console","secrets":1},{"name":"deployer","secrets":1}],"openshift-console-operator":[{"name":"builder","secrets":1},{"name":"console-operator","secrets":1},{"name":"deployer","secrets":1}],"openshift-console-user-settings":[{"name":"builder","secrets":1},{"name":"deployer","secrets":1}],"openshift-controller-manager":[{"name":"builder","secrets":1},{"name":"default","secrets":1},{"name":"deployer","secrets":1},{"name":"openshift-controller-manager-sa","secrets":1}],"openshift-controller-manager-operator":[{"name":"builder","secrets":1},{"name":"default","secrets":1},{"name":"deployer","secrets":1},{"name":"openshift-controller-manager-operator","secrets":1}],"openshift-dns":[{"name":"builder","secrets":1},{"name":"default","secrets":1},{"name":"deployer","secrets":1},{"name":"dns","secrets":1},{"name":"node-resolver","secrets":1}],"openshift-dns-operator":[{"name":"builder","secrets":1},{"name":"default","secrets":1},{"name":"deployer","secrets":1},{"name":"dns-operator","secrets":1}],"openshift-etcd":[{"name":"builder","secrets":1},{"name":"default","secrets":1},{"name":"deployer","secrets":1},{"name":"etcd-backup-sa","secrets":1},{"name":"etcd-sa","secrets":1},{"name":"installer-sa","secrets":1}],"openshift-etcd-operator":[{"name":"builder","secrets":1},{"name":"default","secrets":1},{"name":"deployer","secrets":1},{"name":"etcd-operator","secrets":1}],"openshift-host-network":[{"name":"builder","secrets":1},{"name":"default","secrets":1},{"name":"deployer","secrets":1}],"openshift-image-registry":[{"name":"builder","secrets":1},{"name":"cluster-image-registry-operator","secrets":1},{"name":"default","secrets":1},{"name":"deployer","secrets":1},{"name":"node-ca","secrets":1},{"name":"pruner","secrets":1}],"openshift-infra":[{"name":"build-config-change-controller","secrets":1},{"name":"build-controller","secrets":1},{"name":"builder","secrets":1},{"name":"cluster-csr-approver-controller","secrets":1},{"name":"cluster-quota-reconciliation-controller","secrets":1},{"name":"default","secrets":1},{"name":"default-rolebindings-controller","secrets":1},{"name":"deployer","secrets":1},{"name":"deployer-controller","secrets":1},{"name":"deploymentconfig-controller","secrets":1},{"name":"image-import-controller","secrets":1},{"name":"image-trigger-controller","secrets":1},{"name":"ingress-to-route-controller","secrets":1},{"name":"namespace-security-allocation-controller","secrets":1},{"name":"node-bootstrapper","secrets":1},{"name":"origin-namespace-controller","secrets":1},{"name":"podsecurity-admission-label-syncer-controller","secrets":1},{"name":"privileged-namespaces-psa-label-syncer","secrets":1},{"name":"pv-recycler-controller","secrets":1},{"name":"resourcequota-controller","secrets":1},{"name":"serviceaccount-controller","secrets":1},{"name":"serviceaccount-pull-secrets-controller","secrets":1},{"name":"template-instance-controller","secrets":1},{"name":"template-instance-finalizer-controller","secrets":1},{"name":"unidling-controller","secrets":1}],"openshift-ingress":[{"name":"builder","secrets":1},{"name":"default","secrets":1},{"name":"deployer","secrets":1},{"name":"router","secrets":1}],"openshift-ingress-canary":[{"name":"builder","secrets":1},{"name":"deployer","secrets":1}],"openshift-ingress-operator":[{"name":"builder","secrets":1},{"name":"default","secrets":1},{"name":"deployer","secrets":1},{"name":"ingress-operator","secrets":1}],"openshift-insights":[{"name":"builder","secrets":1},{"name":"default","secrets":1},{"name":"deployer","secrets":1},{"name":"gather","secrets":1},{"name":"operator","secrets":1}],"openshift-kni-infra":[{"name":"builder","secrets":1},{"name":"default","secrets":1},{"name":"deployer","secrets":1}],"openshift-kube-apiserver":[{"name":"builder","secrets":1},{"name":"default","secrets":1},{"name":"deployer","secrets":1},{"name":"installer-sa","secrets":1},{"name":"localhost-recovery-client","secrets":1}],"openshift-kube-apiserver-operator":[{"name":"builder","secrets":1},{"name":"default","secrets":1},{"name":"deployer","secrets":1},{"name":"kube-apiserver-operator","secrets":1}],"openshift-kube-controller-manager":[{"name":"builder","secrets":1},{"name":"default","secrets":1},{"name":"deployer","secrets":1},{"name":"installer-sa","secrets":1},{"name":"kube-controller-manager-sa","secrets":1},{"name":"localhost-recovery-client","secrets":1}],"openshift-kube-controller-manager-operator":[{"name":"builder","secrets":1},{"name":"default","secrets":1},{"name":"deployer","secrets":1},{"name":"kube-controller-manager-operator","secrets":1}],"openshift-kube-scheduler":[{"name":"builder","secrets":1},{"name":"default","secrets":1},{"name":"deployer","secrets":1},{"name":"installer-sa","secrets":1},{"name":"localhost-recovery-client","secrets":1},{"name":"openshift-kube-scheduler-sa","secrets":1}],"openshift-kube-scheduler-operator":[{"name":"builder","secrets":1},{"name":"default","secrets":1},{"name":"deployer","secrets":1},{"name":"openshift-kube-scheduler-operator","secrets":1}],"openshift-kube-storage-version-migrator":[{"name":"builder","secrets":1},{"name":"default","secrets":1},{"name":"deployer","secrets":1},{"name":"kube-storage-version-migrator-sa","secrets":1}],"openshift-kube-storage-version-migrator-operator":[{"name":"builder","secrets":1},{"name":"default","secrets":1},{"name":"deployer","secrets":1},{"name":"kube-storage-version-migrator-operator","secrets":1}],"openshift-machine-api":[{"name":"builder","secrets":1},{"name":"cluster-autoscaler","secrets":1},{"name":"cluster-autoscaler-operator","secrets":1},{"name":"cluster-baremetal-operator","secrets":1},{"name":"control-plane-machine-set-operator","secrets":1},{"name":"default","secrets":1},{"name":"deployer","secrets":1},{"name":"machine-api-controllers","secrets":1},{"name":"machine-api-operator","secrets":1},{"name":"machine-api-termination-handler","secrets":1}],"openshift-machine-config-operator":[{"name":"builder","secrets":1},{"name":"default","secrets":1},{"name":"deployer","secrets":1},{"name":"machine-config-daemon","secrets":1},{"name":"machine-config-operator","secrets":1}],"openshift-marketplace":[{"name":"builder","secrets":1},{"name":"certified-operators","secrets":1},{"name":"community-operators","secrets":1},{"name":"default","secrets":1},{"name":"deployer","secrets":1},{"name":"marketplace-operator","secrets":1},{"name":"redhat-marketplace","secrets":1},{"name":"redhat-operators","secrets":1}],"openshift-monitoring":[{"name":"builder","secrets":1},{"name":"cluster-monitoring-operator","secrets":1},{"name":"default","secrets":1},{"name":"deployer","secrets":1},{"name":"prometheus-operator","secrets":1},{"name":"prometheus-operator-admission-webhook","secrets":1}],"openshift-multus":[{"name":"builder","secrets":1},{"name":"default","secrets":1},{"name":"deployer","secrets":1},{"name":"metrics-daemon-sa","secrets":1},{"name":"multus","secrets":1},{"name":"multus-ac","secrets":1},{"name":"multus-ancillary-tools","secrets":1}],"openshift-network-diagnostics":[{"name":"builder","secrets":1},{"name":"default","secrets":1},{"name":"deployer","secrets":1},{"name":"network-diagnostics","secrets":1}],"openshift-network-node-identity":[{"name":"builder","secrets":1},{"name":"default","secrets":1},{"name":"deployer","secrets":1},{"name":"network-node-identity","secrets":1}],"openshift-network-operator":[{"name":"builder","secrets":1},{"name":"cluster-network-operator","secrets":1},{"name":"default","secrets":1},{"name":"deployer","secrets":1},{"name":"iptables-alerter","secrets":1}],"openshift-node":[{"name":"builder","secrets":1},{"name":"default","secrets":1},{"name":"deployer","secrets":1}],"openshift-nutanix-infra":[{"name":"builder","secrets":1},{"name":"default","secrets":1},{"name":"deployer","secrets":1}],"openshift-oauth-apiserver":[{"name":"builder","secrets":1},{"name":"default","secrets":1},{"name":"deployer","secrets":1},{"name":"oauth-apiserver-sa","secrets":1}],"openshift-openstack-infra":[{"name":"builder","secrets":1},{"name":"default","secrets":1},{"name":"deployer","secrets":1}],"openshift-operator-controller":[{"name":"builder","secrets":1},{"name":"default","secrets":1},{"name":"deployer","secrets":1},{"name":"operator-controller-controller-manager","secrets":1}],"openshift-operator-lifecycle-manager":[{"name":"builder","secrets":1},{"name":"collect-profiles","secrets":1},{"name":"default","secrets":1},{"name":"deployer","secrets":1},{"name":"olm-operator-serviceaccount","secrets":1}],"openshift-operators":[{"name":"builder","secrets":1},{"name":"default","secrets":1},{"name":"deployer","secrets":1}],"openshift-ovirt-infra":[{"name":"builder","secrets":1},{"name":"default","secrets":1},{"name":"deployer","secrets":1}],"openshift-ovn-kubernetes":[{"name":"builder","secrets":1},{"name":"default","secrets":1},{"name":"deployer","secrets":1},{"name":"ovn-kubernetes-control-plane","secrets":1},{"name":"ovn-kubernetes-node","secrets":1}],"openshift-route-controller-manager":[{"name":"builder","secrets":1},{"name":"default","secrets":1},{"name":"deployer","secrets":1},{"name":"route-controller-manager-sa","secrets":1}],"openshift-service-ca":[{"name":"builder","secrets":1},{"name":"default","secrets":1},{"name":"deployer","secrets":1},{"name":"service-ca","secrets":1}],"openshift-service-ca-operator":[{"name":"builder","secrets":1},{"name":"default","secrets":1},{"name":"deployer","secrets":1},{"name":"service-ca-operator","secrets":1}],"openshift-user-workload-monitoring":[{"name":"builder","secrets":1},{"name":"default","secrets":1},{"name":"deployer","secrets":1}],"openshift-vsphere-infra":[{"name":"builder","secrets":1},{"name":"default","secrets":1},{"name":"deployer","secrets":1}]}}}config/installplans.json0000640000000000000000000000010015114274074014100 0ustar0000000000000000{"items":null,"stats":{"TOTAL_COUNT":0,"TOTAL_NONUNIQ_COUNT":0}}config/node/logs/master-0.log0000640000000000000000000005505615114274061014545 0ustar0000000000000000Dec 04 11:59:27.840778 master-0 kubenswrapper[35367]: E1204 11:59:27.840720 35367 secret.go:189] Couldn't get secret openshift-machine-api/cluster-baremetal-webhook-server-cert: failed to sync secret cache: timed out waiting for the condition Dec 04 11:59:27.840909 master-0 kubenswrapper[35367]: E1204 11:59:27.840883 35367 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/7e3160a9-11d1-4845-ba30-1a49ae7339a9-cert podName:7e3160a9-11d1-4845-ba30-1a49ae7339a9 nodeName:}" failed. No retries permitted until 2025-12-04 11:59:28.340850877 +0000 UTC m=+11.778912135 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/7e3160a9-11d1-4845-ba30-1a49ae7339a9-cert") pod "cluster-baremetal-operator-78f758c7b9-zgkh5" (UID: "7e3160a9-11d1-4845-ba30-1a49ae7339a9") : failed to sync secret cache: timed out waiting for the condition Dec 04 11:59:27.840993 master-0 kubenswrapper[35367]: E1204 11:59:27.840941 35367 configmap.go:193] Couldn't get configMap openshift-machine-config-operator/kube-rbac-proxy: failed to sync configmap cache: timed out waiting for the condition Dec 04 11:59:27.841071 master-0 kubenswrapper[35367]: E1204 11:59:27.841048 35367 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/cdb50850-779d-4179-ba1d-65ffeb4bb2e9-auth-proxy-config podName:cdb50850-779d-4179-ba1d-65ffeb4bb2e9 nodeName:}" failed. No retries permitted until 2025-12-04 11:59:28.341030552 +0000 UTC m=+11.779091810 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "auth-proxy-config" (UniqueName: "kubernetes.io/configmap/cdb50850-779d-4179-ba1d-65ffeb4bb2e9-auth-proxy-config") pod "machine-config-operator-dc5d7666f-p2cmn" (UID: "cdb50850-779d-4179-ba1d-65ffeb4bb2e9") : failed to sync configmap cache: timed out waiting for the condition Dec 04 11:59:27.841200 master-0 kubenswrapper[35367]: E1204 11:59:27.841177 35367 configmap.go:193] Couldn't get configMap openshift-machine-api/baremetal-kube-rbac-proxy: failed to sync configmap cache: timed out waiting for the condition Dec 04 11:59:27.841320 master-0 kubenswrapper[35367]: E1204 11:59:27.841307 35367 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/7e3160a9-11d1-4845-ba30-1a49ae7339a9-config podName:7e3160a9-11d1-4845-ba30-1a49ae7339a9 nodeName:}" failed. No retries permitted until 2025-12-04 11:59:28.341296991 +0000 UTC m=+11.779358249 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config" (UniqueName: "kubernetes.io/configmap/7e3160a9-11d1-4845-ba30-1a49ae7339a9-config") pod "cluster-baremetal-operator-78f758c7b9-zgkh5" (UID: "7e3160a9-11d1-4845-ba30-1a49ae7339a9") : failed to sync configmap cache: timed out waiting for the condition Dec 04 11:59:27.841897 master-0 kubenswrapper[35367]: E1204 11:59:27.841878 35367 configmap.go:193] Couldn't get configMap openshift-cloud-credential-operator/cco-trusted-ca: failed to sync configmap cache: timed out waiting for the condition Dec 04 11:59:27.842055 master-0 kubenswrapper[35367]: E1204 11:59:27.841879 35367 configmap.go:193] Couldn't get configMap openshift-cloud-controller-manager-operator/cloud-controller-manager-images: failed to sync configmap cache: timed out waiting for the condition Dec 04 11:59:27.842055 master-0 kubenswrapper[35367]: E1204 11:59:27.842035 35367 configmap.go:193] Couldn't get configMap openshift-insights/trusted-ca-bundle: failed to sync configmap cache: timed out waiting for the condition Dec 04 11:59:27.842173 master-0 kubenswrapper[35367]: E1204 11:59:27.842011 35367 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/64d30b80-2336-49c2-a2c2-1e0694036a84-cco-trusted-ca podName:64d30b80-2336-49c2-a2c2-1e0694036a84 nodeName:}" failed. No retries permitted until 2025-12-04 11:59:28.341999863 +0000 UTC m=+11.780061311 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cco-trusted-ca" (UniqueName: "kubernetes.io/configmap/64d30b80-2336-49c2-a2c2-1e0694036a84-cco-trusted-ca") pod "cloud-credential-operator-698c598cfc-95jdn" (UID: "64d30b80-2336-49c2-a2c2-1e0694036a84") : failed to sync configmap cache: timed out waiting for the condition Dec 04 11:59:27.842173 master-0 kubenswrapper[35367]: E1204 11:59:27.842144 35367 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/f79f39e0-4aa6-4eea-b8f5-7b5ef18e8d48-images podName:f79f39e0-4aa6-4eea-b8f5-7b5ef18e8d48 nodeName:}" failed. No retries permitted until 2025-12-04 11:59:28.342119867 +0000 UTC m=+11.780181295 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "images" (UniqueName: "kubernetes.io/configmap/f79f39e0-4aa6-4eea-b8f5-7b5ef18e8d48-images") pod "cluster-cloud-controller-manager-operator-74f484689c-jmfn2" (UID: "f79f39e0-4aa6-4eea-b8f5-7b5ef18e8d48") : failed to sync configmap cache: timed out waiting for the condition Dec 04 11:59:27.842296 master-0 kubenswrapper[35367]: E1204 11:59:27.842176 35367 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/b668f740-dbc3-48e0-9ebc-83bae3a43c83-trusted-ca-bundle podName:b668f740-dbc3-48e0-9ebc-83bae3a43c83 nodeName:}" failed. No retries permitted until 2025-12-04 11:59:28.342161788 +0000 UTC m=+11.780223276 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "trusted-ca-bundle" (UniqueName: "kubernetes.io/configmap/b668f740-dbc3-48e0-9ebc-83bae3a43c83-trusted-ca-bundle") pod "insights-operator-55965856b6-skbmb" (UID: "b668f740-dbc3-48e0-9ebc-83bae3a43c83") : failed to sync configmap cache: timed out waiting for the condition Dec 04 11:59:27.843430 master-0 kubenswrapper[35367]: E1204 11:59:27.843404 35367 configmap.go:193] Couldn't get configMap openshift-insights/service-ca-bundle: failed to sync configmap cache: timed out waiting for the condition Dec 04 11:59:27.843501 master-0 kubenswrapper[35367]: E1204 11:59:27.843456 35367 secret.go:189] Couldn't get secret openshift-machine-api/cluster-autoscaler-operator-cert: failed to sync secret cache: timed out waiting for the condition Dec 04 11:59:27.843548 master-0 kubenswrapper[35367]: E1204 11:59:27.843477 35367 secret.go:189] Couldn't get secret openshift-cloud-controller-manager-operator/cloud-controller-manager-operator-tls: failed to sync secret cache: timed out waiting for the condition Dec 04 11:59:27.843548 master-0 kubenswrapper[35367]: E1204 11:59:27.843461 35367 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/b668f740-dbc3-48e0-9ebc-83bae3a43c83-service-ca-bundle podName:b668f740-dbc3-48e0-9ebc-83bae3a43c83 nodeName:}" failed. No retries permitted until 2025-12-04 11:59:28.343449449 +0000 UTC m=+11.781510877 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "service-ca-bundle" (UniqueName: "kubernetes.io/configmap/b668f740-dbc3-48e0-9ebc-83bae3a43c83-service-ca-bundle") pod "insights-operator-55965856b6-skbmb" (UID: "b668f740-dbc3-48e0-9ebc-83bae3a43c83") : failed to sync configmap cache: timed out waiting for the condition Dec 04 11:59:27.843548 master-0 kubenswrapper[35367]: E1204 11:59:27.843538 35367 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e7fc7c16-5bca-49e5-aff0-7a8f80c6b639-cert podName:e7fc7c16-5bca-49e5-aff0-7a8f80c6b639 nodeName:}" failed. No retries permitted until 2025-12-04 11:59:28.343524442 +0000 UTC m=+11.781585700 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/e7fc7c16-5bca-49e5-aff0-7a8f80c6b639-cert") pod "cluster-autoscaler-operator-5f49d774cd-894dk" (UID: "e7fc7c16-5bca-49e5-aff0-7a8f80c6b639") : failed to sync secret cache: timed out waiting for the condition Dec 04 11:59:27.843672 master-0 kubenswrapper[35367]: E1204 11:59:27.843556 35367 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f79f39e0-4aa6-4eea-b8f5-7b5ef18e8d48-cloud-controller-manager-operator-tls podName:f79f39e0-4aa6-4eea-b8f5-7b5ef18e8d48 nodeName:}" failed. No retries permitted until 2025-12-04 11:59:28.343546933 +0000 UTC m=+11.781608191 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cloud-controller-manager-operator-tls" (UniqueName: "kubernetes.io/secret/f79f39e0-4aa6-4eea-b8f5-7b5ef18e8d48-cloud-controller-manager-operator-tls") pod "cluster-cloud-controller-manager-operator-74f484689c-jmfn2" (UID: "f79f39e0-4aa6-4eea-b8f5-7b5ef18e8d48") : failed to sync secret cache: timed out waiting for the condition Dec 04 11:59:27.843747 master-0 kubenswrapper[35367]: E1204 11:59:27.843727 35367 configmap.go:193] Couldn't get configMap openshift-controller-manager/client-ca: failed to sync configmap cache: timed out waiting for the condition Dec 04 11:59:27.843876 master-0 kubenswrapper[35367]: E1204 11:59:27.843862 35367 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/24506aa4-ab78-49df-bb58-59093498f13d-client-ca podName:24506aa4-ab78-49df-bb58-59093498f13d nodeName:}" failed. No retries permitted until 2025-12-04 11:59:28.343849191 +0000 UTC m=+11.781910639 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "client-ca" (UniqueName: "kubernetes.io/configmap/24506aa4-ab78-49df-bb58-59093498f13d-client-ca") pod "controller-manager-6686654b8d-rrndk" (UID: "24506aa4-ab78-49df-bb58-59093498f13d") : failed to sync configmap cache: timed out waiting for the condition Dec 04 11:59:27.844669 master-0 kubenswrapper[35367]: E1204 11:59:27.844643 35367 configmap.go:193] Couldn't get configMap openshift-controller-manager/openshift-global-ca: failed to sync configmap cache: timed out waiting for the condition Dec 04 11:59:27.844743 master-0 kubenswrapper[35367]: E1204 11:59:27.844699 35367 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/24506aa4-ab78-49df-bb58-59093498f13d-proxy-ca-bundles podName:24506aa4-ab78-49df-bb58-59093498f13d nodeName:}" failed. No retries permitted until 2025-12-04 11:59:28.344687858 +0000 UTC m=+11.782749116 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "proxy-ca-bundles" (UniqueName: "kubernetes.io/configmap/24506aa4-ab78-49df-bb58-59093498f13d-proxy-ca-bundles") pod "controller-manager-6686654b8d-rrndk" (UID: "24506aa4-ab78-49df-bb58-59093498f13d") : failed to sync configmap cache: timed out waiting for the condition Dec 04 11:59:27.845032 master-0 kubenswrapper[35367]: E1204 11:59:27.845016 35367 configmap.go:193] Couldn't get configMap openshift-machine-api/cluster-baremetal-operator-images: failed to sync configmap cache: timed out waiting for the condition Dec 04 11:59:27.845132 master-0 kubenswrapper[35367]: E1204 11:59:27.845120 35367 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/7e3160a9-11d1-4845-ba30-1a49ae7339a9-images podName:7e3160a9-11d1-4845-ba30-1a49ae7339a9 nodeName:}" failed. No retries permitted until 2025-12-04 11:59:28.345109982 +0000 UTC m=+11.783171240 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "images" (UniqueName: "kubernetes.io/configmap/7e3160a9-11d1-4845-ba30-1a49ae7339a9-images") pod "cluster-baremetal-operator-78f758c7b9-zgkh5" (UID: "7e3160a9-11d1-4845-ba30-1a49ae7339a9") : failed to sync configmap cache: timed out waiting for the condition Dec 04 11:59:27.845745 master-0 kubenswrapper[35367]: E1204 11:59:27.845709 35367 secret.go:189] Couldn't get secret openshift-machine-api/machine-api-operator-tls: failed to sync secret cache: timed out waiting for the condition Dec 04 11:59:27.845868 master-0 kubenswrapper[35367]: E1204 11:59:27.845844 35367 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c50317d3-f7cd-4133-845e-44add57ac378-machine-api-operator-tls podName:c50317d3-f7cd-4133-845e-44add57ac378 nodeName:}" failed. No retries permitted until 2025-12-04 11:59:28.345790224 +0000 UTC m=+11.783851482 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "machine-api-operator-tls" (UniqueName: "kubernetes.io/secret/c50317d3-f7cd-4133-845e-44add57ac378-machine-api-operator-tls") pod "machine-api-operator-88d48b57d-9fjtd" (UID: "c50317d3-f7cd-4133-845e-44add57ac378") : failed to sync secret cache: timed out waiting for the condition Dec 04 11:59:27.845942 master-0 kubenswrapper[35367]: E1204 11:59:27.845877 35367 secret.go:189] Couldn't get secret openshift-machine-api/cluster-baremetal-operator-tls: failed to sync secret cache: timed out waiting for the condition Dec 04 11:59:27.845942 master-0 kubenswrapper[35367]: E1204 11:59:27.845938 35367 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/7e3160a9-11d1-4845-ba30-1a49ae7339a9-cluster-baremetal-operator-tls podName:7e3160a9-11d1-4845-ba30-1a49ae7339a9 nodeName:}" failed. No retries permitted until 2025-12-04 11:59:28.345925718 +0000 UTC m=+11.783987156 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cluster-baremetal-operator-tls" (UniqueName: "kubernetes.io/secret/7e3160a9-11d1-4845-ba30-1a49ae7339a9-cluster-baremetal-operator-tls") pod "cluster-baremetal-operator-78f758c7b9-zgkh5" (UID: "7e3160a9-11d1-4845-ba30-1a49ae7339a9") : failed to sync secret cache: timed out waiting for the condition Dec 04 11:59:27.847842 master-0 kubenswrapper[35367]: E1204 11:59:27.846924 35367 configmap.go:193] Couldn't get configMap openshift-route-controller-manager/config: failed to sync configmap cache: timed out waiting for the condition Dec 04 11:59:27.847842 master-0 kubenswrapper[35367]: E1204 11:59:27.846996 35367 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/e4d7939a-5961-4608-b910-73e71aa55bf6-config podName:e4d7939a-5961-4608-b910-73e71aa55bf6 nodeName:}" failed. No retries permitted until 2025-12-04 11:59:28.346980931 +0000 UTC m=+11.785042359 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config" (UniqueName: "kubernetes.io/configmap/e4d7939a-5961-4608-b910-73e71aa55bf6-config") pod "route-controller-manager-95cb5f987-46bsk" (UID: "e4d7939a-5961-4608-b910-73e71aa55bf6") : failed to sync configmap cache: timed out waiting for the condition Dec 04 11:59:27.848940 master-0 kubenswrapper[35367]: E1204 11:59:27.848845 35367 configmap.go:193] Couldn't get configMap openshift-machine-api/machine-api-operator-images: failed to sync configmap cache: timed out waiting for the condition Dec 04 11:59:27.848940 master-0 kubenswrapper[35367]: E1204 11:59:27.848922 35367 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/c50317d3-f7cd-4133-845e-44add57ac378-images podName:c50317d3-f7cd-4133-845e-44add57ac378 nodeName:}" failed. No retries permitted until 2025-12-04 11:59:28.348901922 +0000 UTC m=+11.786963180 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "images" (UniqueName: "kubernetes.io/configmap/c50317d3-f7cd-4133-845e-44add57ac378-images") pod "machine-api-operator-88d48b57d-9fjtd" (UID: "c50317d3-f7cd-4133-845e-44add57ac378") : failed to sync configmap cache: timed out waiting for the condition Dec 04 11:59:27.851220 master-0 kubenswrapper[35367]: E1204 11:59:27.851194 35367 configmap.go:193] Couldn't get configMap openshift-machine-api/kube-rbac-proxy: failed to sync configmap cache: timed out waiting for the condition Dec 04 11:59:27.851353 master-0 kubenswrapper[35367]: E1204 11:59:27.851256 35367 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/c50317d3-f7cd-4133-845e-44add57ac378-config podName:c50317d3-f7cd-4133-845e-44add57ac378 nodeName:}" failed. No retries permitted until 2025-12-04 11:59:28.351244966 +0000 UTC m=+11.789306224 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config" (UniqueName: "kubernetes.io/configmap/c50317d3-f7cd-4133-845e-44add57ac378-config") pod "machine-api-operator-88d48b57d-9fjtd" (UID: "c50317d3-f7cd-4133-845e-44add57ac378") : failed to sync configmap cache: timed out waiting for the condition Dec 04 11:59:27.851353 master-0 kubenswrapper[35367]: E1204 11:59:27.851294 35367 secret.go:189] Couldn't get secret openshift-controller-manager/serving-cert: failed to sync secret cache: timed out waiting for the condition Dec 04 11:59:27.851353 master-0 kubenswrapper[35367]: E1204 11:59:27.851326 35367 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/24506aa4-ab78-49df-bb58-59093498f13d-serving-cert podName:24506aa4-ab78-49df-bb58-59093498f13d nodeName:}" failed. No retries permitted until 2025-12-04 11:59:28.351320319 +0000 UTC m=+11.789381577 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "serving-cert" (UniqueName: "kubernetes.io/secret/24506aa4-ab78-49df-bb58-59093498f13d-serving-cert") pod "controller-manager-6686654b8d-rrndk" (UID: "24506aa4-ab78-49df-bb58-59093498f13d") : failed to sync secret cache: timed out waiting for the condition Dec 04 11:59:27.851753 master-0 kubenswrapper[35367]: E1204 11:59:27.851726 35367 configmap.go:193] Couldn't get configMap openshift-controller-manager/config: failed to sync configmap cache: timed out waiting for the condition Dec 04 11:59:27.851797 master-0 kubenswrapper[35367]: E1204 11:59:27.851778 35367 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/24506aa4-ab78-49df-bb58-59093498f13d-config podName:24506aa4-ab78-49df-bb58-59093498f13d nodeName:}" failed. No retries permitted until 2025-12-04 11:59:28.351768503 +0000 UTC m=+11.789829761 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config" (UniqueName: "kubernetes.io/configmap/24506aa4-ab78-49df-bb58-59093498f13d-config") pod "controller-manager-6686654b8d-rrndk" (UID: "24506aa4-ab78-49df-bb58-59093498f13d") : failed to sync configmap cache: timed out waiting for the condition Dec 04 11:59:27.891079 master-0 kubenswrapper[35367]: E1204 11:59:27.890196 35367 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd75143d9bc4a2dc15781dc51ccff632a.slice/crio-59191695c89f8cecbba881756b5ffc5e4c892fd02facf20cc198ade1bf4ec267\": RecentStats: unable to find data in memory cache]" Dec 04 11:59:28.842566 master-0 kubenswrapper[35367]: E1204 11:59:28.842502 35367 kubelet.go:2526] "Housekeeping took longer than expected" err="housekeeping took too long" expected="1s" actual="2.08s" Dec 04 11:59:28.892147 master-0 kubenswrapper[35367]: E1204 11:59:28.892105 35367 projected.go:288] Couldn't get configMap openshift-kube-apiserver/kube-root-ca.crt: object "openshift-kube-apiserver"/"kube-root-ca.crt" not registered Dec 04 11:59:28.892331 master-0 kubenswrapper[35367]: E1204 11:59:28.892318 35367 projected.go:194] Error preparing data for projected volume kube-api-access for pod openshift-kube-apiserver/installer-3-master-0: object "openshift-kube-apiserver"/"kube-root-ca.crt" not registered Dec 04 11:59:28.892465 master-0 kubenswrapper[35367]: E1204 11:59:28.892454 35367 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/0fad065c-94c9-4a38-ab76-cf0266bdc04b-kube-api-access podName:0fad065c-94c9-4a38-ab76-cf0266bdc04b nodeName:}" failed. No retries permitted until 2025-12-04 11:59:29.392429702 +0000 UTC m=+12.830490960 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access" (UniqueName: "kubernetes.io/projected/0fad065c-94c9-4a38-ab76-cf0266bdc04b-kube-api-access") pod "installer-3-master-0" (UID: "0fad065c-94c9-4a38-ab76-cf0266bdc04b") : object "openshift-kube-apiserver"/"kube-root-ca.crt" not registered Dec 04 11:59:30.133996 master-0 kubenswrapper[35367]: E1204 11:59:30.133952 35367 kubelet.go:2526] "Housekeeping took longer than expected" err="housekeeping took too long" expected="1s" actual="1.291s" Dec 04 11:59:42.257124 master-0 kubenswrapper[35367]: E1204 11:59:42.256671 35367 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b02fad9634552ca9f9ab767de0ddd976a9b715f9ba49fca68b0847641cd92467\": container with ID starting with b02fad9634552ca9f9ab767de0ddd976a9b715f9ba49fca68b0847641cd92467 not found: ID does not exist" containerID="b02fad9634552ca9f9ab767de0ddd976a9b715f9ba49fca68b0847641cd92467" Dec 04 12:00:44.801016 master-0 kubenswrapper[35367]: E1204 12:00:44.800948 35367 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cluster-policy-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=cluster-policy-controller pod=kube-controller-manager-master-0_openshift-kube-controller-manager(e6b437c60bb18680f4492b00b294e872)\"" pod="openshift-kube-controller-manager/kube-controller-manager-master-0" podUID="e6b437c60bb18680f4492b00b294e872" Dec 04 12:00:45.700555 master-0 kubenswrapper[35367]: E1204 12:00:45.700517 35367 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cluster-policy-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=cluster-policy-controller pod=kube-controller-manager-master-0_openshift-kube-controller-manager(e6b437c60bb18680f4492b00b294e872)\"" pod="openshift-kube-controller-manager/kube-controller-manager-master-0" podUID="e6b437c60bb18680f4492b00b294e872" Dec 04 12:00:50.452380 master-0 kubenswrapper[35367]: E1204 12:00:50.452343 35367 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cluster-policy-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=cluster-policy-controller pod=kube-controller-manager-master-0_openshift-kube-controller-manager(e6b437c60bb18680f4492b00b294e872)\"" pod="openshift-kube-controller-manager/kube-controller-manager-master-0" podUID="e6b437c60bb18680f4492b00b294e872" Dec 04 12:01:03.748096 master-0 kubenswrapper[35367]: E1204 12:01:03.747588 35367 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cluster-policy-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=cluster-policy-controller pod=kube-controller-manager-master-0_openshift-kube-controller-manager(e6b437c60bb18680f4492b00b294e872)\"" pod="openshift-kube-controller-manager/kube-controller-manager-master-0" podUID="e6b437c60bb18680f4492b00b294e872" Dec 04 12:01:16.752336 master-0 kubenswrapper[35367]: E1204 12:01:16.751962 35367 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cluster-policy-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=cluster-policy-controller pod=kube-controller-manager-master-0_openshift-kube-controller-manager(e6b437c60bb18680f4492b00b294e872)\"" pod="openshift-kube-controller-manager/kube-controller-manager-master-0" podUID="e6b437c60bb18680f4492b00b294e872"config/clusteroperator/storage.json0000640000000000000000000000346715114274061016314 0ustar0000000000000000{"kind":"ClusterOperator","apiVersion":"config.openshift.io/v1","metadata":{"name":"storage","uid":"102871f4-555b-49a3-844b-9e204a9c8364","resourceVersion":"10307","generation":1,"creationTimestamp":"2025-12-04T11:31:14Z","annotations":{"capability.openshift.io/name":"Storage","include.release.openshift.io/hypershift":"true","include.release.openshift.io/ibm-cloud-managed":"true","include.release.openshift.io/self-managed-high-availability":"true","include.release.openshift.io/single-node-developer":"true"},"ownerReferences":[{"apiVersion":"config.openshift.io/v1","kind":"ClusterVersion","name":"version","uid":"a151bf9a-21b6-4d33-89ec-0e4f83e8eab6","controller":true}]},"spec":{},"status":{"conditions":[{"type":"Degraded","status":"False","lastTransitionTime":"2025-12-04T11:45:48Z","reason":"AsExpected","message":"All is well"},{"type":"Progressing","status":"False","lastTransitionTime":"2025-12-04T11:45:48Z","reason":"AsExpected","message":"All is well"},{"type":"Available","status":"True","lastTransitionTime":"2025-12-04T11:45:48Z","reason":"AsExpected","message":"DefaultStorageClassControllerAvailable: No default StorageClass for this platform"},{"type":"Upgradeable","status":"True","lastTransitionTime":"2025-12-04T11:45:48Z","reason":"AsExpected","message":"All is well"},{"type":"EvaluationConditionsDetected","status":"Unknown","lastTransitionTime":"2025-12-04T11:45:48Z","reason":"NoData"}],"versions":[{"name":"operator","version":"4.18.29"}],"relatedObjects":[{"group":"","resource":"namespaces","name":"openshift-cluster-storage-operator"},{"group":"","resource":"namespaces","name":"openshift-cluster-csi-drivers"},{"group":"operator.openshift.io","resource":"storages","name":"cluster"},{"group":"rbac.authorization.k8s.io","resource":"clusterrolebindings","name":"cluster-storage-operator-role"}],"extension":null}}config/clusteroperator/service-ca.json0000640000000000000000000000340315114274061016657 0ustar0000000000000000{"kind":"ClusterOperator","apiVersion":"config.openshift.io/v1","metadata":{"name":"service-ca","uid":"281e9a38-44ec-4b75-adde-903b1080cebc","resourceVersion":"4987","generation":1,"creationTimestamp":"2025-12-04T11:31:13Z","annotations":{"include.release.openshift.io/hypershift":"true","include.release.openshift.io/ibm-cloud-managed":"true","include.release.openshift.io/self-managed-high-availability":"true","include.release.openshift.io/single-node-developer":"true"},"ownerReferences":[{"apiVersion":"config.openshift.io/v1","kind":"ClusterVersion","name":"version","uid":"a151bf9a-21b6-4d33-89ec-0e4f83e8eab6","controller":true}]},"spec":{},"status":{"conditions":[{"type":"Degraded","status":"False","lastTransitionTime":"2025-12-04T11:37:45Z","reason":"AsExpected","message":"All is well"},{"type":"Progressing","status":"False","lastTransitionTime":"2025-12-04T11:37:53Z","reason":"AsExpected","message":"Progressing: All service-ca-operator deployments updated"},{"type":"Available","status":"True","lastTransitionTime":"2025-12-04T11:37:47Z","reason":"AsExpected","message":"All is well"},{"type":"Upgradeable","status":"True","lastTransitionTime":"2025-12-04T11:37:47Z","reason":"AsExpected","message":"All is well"},{"type":"EvaluationConditionsDetected","status":"Unknown","lastTransitionTime":"2025-12-04T11:37:44Z","reason":"NoData"}],"versions":[{"name":"operator","version":"4.18.29"}],"relatedObjects":[{"group":"operator.openshift.io","resource":"servicecas","name":"cluster"},{"group":"","resource":"namespaces","name":"openshift-config"},{"group":"","resource":"namespaces","name":"openshift-config-managed"},{"group":"","resource":"namespaces","name":"openshift-service-ca-operator"},{"group":"","resource":"namespaces","name":"openshift-service-ca"}],"extension":null}}config/clusteroperator/operator-lifecycle-manager-packageserver.json0000640000000000000000000000305215114274061024656 0ustar0000000000000000{"kind":"ClusterOperator","apiVersion":"config.openshift.io/v1","metadata":{"name":"operator-lifecycle-manager-packageserver","uid":"f651b646-a3e6-48d5-9931-718619a73c02","resourceVersion":"14167","generation":1,"creationTimestamp":"2025-12-04T11:31:15Z","annotations":{"capability.openshift.io/name":"OperatorLifecycleManager","include.release.openshift.io/hypershift":"true","include.release.openshift.io/ibm-cloud-managed":"true","include.release.openshift.io/self-managed-high-availability":"true"},"ownerReferences":[{"apiVersion":"config.openshift.io/v1","kind":"ClusterVersion","name":"version","uid":"a151bf9a-21b6-4d33-89ec-0e4f83e8eab6","controller":true}]},"spec":{},"status":{"conditions":[{"type":"Degraded","status":"False","lastTransitionTime":"2025-12-04T11:38:33Z"},{"type":"Available","status":"False","lastTransitionTime":"2025-12-04T11:38:33Z","reason":"ClusterServiceVersionNotSucceeded","message":"ClusterServiceVersion openshift-operator-lifecycle-manager/packageserver observed in phase Failed with reason: InstallCheckFailed, message: install timeout"},{"type":"Progressing","status":"True","lastTransitionTime":"2025-12-04T12:00:59Z","message":"Working toward 0.0.1-snapshot"},{"type":"Upgradeable","status":"True","lastTransitionTime":"2025-12-04T11:38:34Z","message":"Safe to upgrade"}],"relatedObjects":[{"group":"","resource":"namespaces","name":"openshift-operator-lifecycle-manager"},{"group":"operators.coreos.com","resource":"clusterserviceversions","namespace":"openshift-operator-lifecycle-manager","name":"packageserver"}],"extension":null}}config/clusteroperator/operator-lifecycle-manager-catalog.json0000640000000000000000000000255615114274061023456 0ustar0000000000000000{"kind":"ClusterOperator","apiVersion":"config.openshift.io/v1","metadata":{"name":"operator-lifecycle-manager-catalog","uid":"f22eba84-fef3-488d-b7c5-c26874042b47","resourceVersion":"10325","generation":1,"creationTimestamp":"2025-12-04T11:31:15Z","annotations":{"capability.openshift.io/name":"OperatorLifecycleManager","include.release.openshift.io/hypershift":"true","include.release.openshift.io/ibm-cloud-managed":"true","include.release.openshift.io/self-managed-high-availability":"true"},"ownerReferences":[{"apiVersion":"config.openshift.io/v1","kind":"ClusterVersion","name":"version","uid":"a151bf9a-21b6-4d33-89ec-0e4f83e8eab6","controller":true}]},"spec":{},"status":{"conditions":[{"type":"Degraded","status":"False","lastTransitionTime":"2025-12-04T11:38:33Z"},{"type":"Progressing","status":"False","lastTransitionTime":"2025-12-04T11:45:52Z","message":"Deployed 4.18.0-202511181540.p2.g357bb13.assembly.stream.el9-357bb13"},{"type":"Available","status":"True","lastTransitionTime":"2025-12-04T11:38:33Z"},{"type":"Upgradeable","status":"True","lastTransitionTime":"2025-12-04T11:38:33Z"}],"versions":[{"name":"operator","version":"4.18.29"},{"name":"operator-lifecycle-manager","version":"4.18.0-202511181540.p2.g357bb13.assembly.stream.el9-357bb13"}],"relatedObjects":[{"group":"","resource":"namespaces","name":"openshift-operator-lifecycle-manager"}],"extension":null}}config/clusteroperator/operator-lifecycle-manager.json0000640000000000000000000000264115114274061022041 0ustar0000000000000000{"kind":"ClusterOperator","apiVersion":"config.openshift.io/v1","metadata":{"name":"operator-lifecycle-manager","uid":"e0b9c929-f091-45bf-92b4-3eaafe132732","resourceVersion":"7129","generation":1,"creationTimestamp":"2025-12-04T11:31:13Z","annotations":{"capability.openshift.io/name":"OperatorLifecycleManager","include.release.openshift.io/hypershift":"true","include.release.openshift.io/ibm-cloud-managed":"true","include.release.openshift.io/self-managed-high-availability":"true"},"ownerReferences":[{"apiVersion":"config.openshift.io/v1","kind":"ClusterVersion","name":"version","uid":"a151bf9a-21b6-4d33-89ec-0e4f83e8eab6","controller":true}]},"spec":{},"status":{"conditions":[{"type":"Progressing","status":"False","lastTransitionTime":"2025-12-04T11:38:33Z","message":"Deployed 4.18.0-202511181540.p2.g357bb13.assembly.stream.el9-357bb13"},{"type":"Available","status":"True","lastTransitionTime":"2025-12-04T11:38:33Z"},{"type":"Degraded","status":"False","lastTransitionTime":"2025-12-04T11:38:33Z"},{"type":"Upgradeable","status":"True","lastTransitionTime":"2025-12-04T11:38:33Z"}],"versions":[{"name":"operator","version":"4.18.29"},{"name":"operator-lifecycle-manager","version":"4.18.0-202511181540.p2.g357bb13.assembly.stream.el9-357bb13"}],"relatedObjects":[{"group":"operators.coreos.com","resource":"clusterserviceversions","namespace":"openshift-operator-lifecycle-manager","name":"packageserver"}],"extension":null}}config/clusteroperator/openshift-samples.json0000640000000000000000000000260515114274061020302 0ustar0000000000000000{"kind":"ClusterOperator","apiVersion":"config.openshift.io/v1","metadata":{"name":"openshift-samples","uid":"9a954eb8-27db-4b1b-bba9-fea8c9ef093a","resourceVersion":"14120","generation":1,"creationTimestamp":"2025-12-04T11:31:13Z","annotations":{"capability.openshift.io/name":"openshift-samples","include.release.openshift.io/ibm-cloud-managed":"true","include.release.openshift.io/self-managed-high-availability":"true"},"ownerReferences":[{"apiVersion":"config.openshift.io/v1","kind":"ClusterVersion","name":"version","uid":"a151bf9a-21b6-4d33-89ec-0e4f83e8eab6","controller":true}]},"spec":{},"status":{"conditions":[{"type":"Degraded","status":"False","lastTransitionTime":"2025-12-04T12:00:28Z"},{"type":"Available","status":"True","lastTransitionTime":"2025-12-04T12:00:49Z","message":"Samples installation successful at 4.18.29"},{"type":"Progressing","status":"False","lastTransitionTime":"2025-12-04T12:00:28Z","message":"Samples installation successful at 4.18.29"}],"versions":[{"name":"operator","version":"4.18.29"}],"relatedObjects":[{"group":"samples.operator.openshift.io","resource":"configs","name":"cluster"},{"group":"","resource":"namespaces","name":"openshift-cluster-samples-operator"},{"group":"template.openshift.io","resource":"templates","namespace":"openshift","name":""},{"group":"image.openshift.io","resource":"imagestreams","namespace":"openshift","name":""}],"extension":null}}config/clusteroperator/openshift-controller-manager.json0000640000000000000000000000437215114274061022434 0ustar0000000000000000{"kind":"ClusterOperator","apiVersion":"config.openshift.io/v1","metadata":{"name":"openshift-controller-manager","uid":"14fe6dba-43aa-4871-a35e-e03e67f0bb54","resourceVersion":"14115","generation":1,"creationTimestamp":"2025-12-04T11:31:13Z","annotations":{"exclude.release.openshift.io/internal-openshift-hosted":"true","include.release.openshift.io/self-managed-high-availability":"true","include.release.openshift.io/single-node-developer":"true"},"ownerReferences":[{"apiVersion":"config.openshift.io/v1","kind":"ClusterVersion","name":"version","uid":"a151bf9a-21b6-4d33-89ec-0e4f83e8eab6","controller":true}]},"spec":{},"status":{"conditions":[{"type":"Degraded","status":"False","lastTransitionTime":"2025-12-04T11:37:45Z","reason":"AsExpected","message":"All is well"},{"type":"Progressing","status":"True","lastTransitionTime":"2025-12-04T11:37:49Z","reason":"_DesiredStateNotYetAchieved","message":"Progressing: deployment/controller-manager: observed generation is 5, desired generation is 6.\nProgressing: deployment/controller-manager: available replicas is 0, desired available replicas \u003e 1\nProgressing: deployment/route-controller-manager: observed generation is 5, desired generation is 6.\nProgressing: deployment/route-controller-manager: available replicas is 0, desired available replicas \u003e 1"},{"type":"Available","status":"False","lastTransitionTime":"2025-12-04T11:50:40Z","reason":"_NoPodsAvailable","message":"Available: no pods available on any node."},{"type":"Upgradeable","status":"True","lastTransitionTime":"2025-12-04T11:37:49Z","reason":"AsExpected","message":"All is well"},{"type":"EvaluationConditionsDetected","status":"Unknown","lastTransitionTime":"2025-12-04T11:37:44Z","reason":"NoData"}],"versions":[{"name":"operator","version":"4.18.29"}],"relatedObjects":[{"group":"operator.openshift.io","resource":"openshiftcontrollermanagers","name":"cluster"},{"group":"","resource":"namespaces","name":"openshift-config"},{"group":"","resource":"namespaces","name":"openshift-config-managed"},{"group":"","resource":"namespaces","name":"openshift-controller-manager-operator"},{"group":"","resource":"namespaces","name":"openshift-controller-manager"},{"group":"","resource":"namespaces","name":"openshift-route-controller-manager"}],"extension":null}}config/clusteroperator/openshift-apiserver.json0000640000000000000000000000552415114274061020641 0ustar0000000000000000{"kind":"ClusterOperator","apiVersion":"config.openshift.io/v1","metadata":{"name":"openshift-apiserver","uid":"dee26556-64bf-4b7a-b694-120c9e499e55","resourceVersion":"12112","generation":1,"creationTimestamp":"2025-12-04T11:31:13Z","annotations":{"exclude.release.openshift.io/internal-openshift-hosted":"true","include.release.openshift.io/self-managed-high-availability":"true","include.release.openshift.io/single-node-developer":"true"},"ownerReferences":[{"apiVersion":"config.openshift.io/v1","kind":"ClusterVersion","name":"version","uid":"a151bf9a-21b6-4d33-89ec-0e4f83e8eab6","controller":true}]},"spec":{},"status":{"conditions":[{"type":"Degraded","status":"False","lastTransitionTime":"2025-12-04T11:37:54Z","reason":"AsExpected","message":"All is well"},{"type":"Progressing","status":"False","lastTransitionTime":"2025-12-04T11:38:42Z","reason":"AsExpected","message":"All is well"},{"type":"Available","status":"True","lastTransitionTime":"2025-12-04T11:59:55Z","reason":"AsExpected","message":"All is well"},{"type":"Upgradeable","status":"True","lastTransitionTime":"2025-12-04T11:37:53Z","reason":"AsExpected","message":"All is well"},{"type":"EvaluationConditionsDetected","status":"Unknown","lastTransitionTime":"2025-12-04T11:37:53Z","reason":"NoData"}],"versions":[{"name":"operator","version":"4.18.29"},{"name":"openshift-apiserver","version":"4.18.29"}],"relatedObjects":[{"group":"operator.openshift.io","resource":"openshiftapiservers","name":"cluster"},{"group":"","resource":"namespaces","name":"openshift-config"},{"group":"","resource":"namespaces","name":"openshift-config-managed"},{"group":"","resource":"namespaces","name":"openshift-apiserver-operator"},{"group":"","resource":"namespaces","name":"openshift-apiserver"},{"group":"","resource":"namespaces","name":"openshift-etcd-operator"},{"group":"","resource":"endpoints","namespace":"openshift-etcd","name":"host-etcd-2"},{"group":"controlplane.operator.openshift.io","resource":"podnetworkconnectivitychecks","namespace":"openshift-apiserver","name":""},{"group":"apiregistration.k8s.io","resource":"apiservices","name":"v1.apps.openshift.io"},{"group":"apiregistration.k8s.io","resource":"apiservices","name":"v1.authorization.openshift.io"},{"group":"apiregistration.k8s.io","resource":"apiservices","name":"v1.build.openshift.io"},{"group":"apiregistration.k8s.io","resource":"apiservices","name":"v1.image.openshift.io"},{"group":"apiregistration.k8s.io","resource":"apiservices","name":"v1.project.openshift.io"},{"group":"apiregistration.k8s.io","resource":"apiservices","name":"v1.quota.openshift.io"},{"group":"apiregistration.k8s.io","resource":"apiservices","name":"v1.route.openshift.io"},{"group":"apiregistration.k8s.io","resource":"apiservices","name":"v1.security.openshift.io"},{"group":"apiregistration.k8s.io","resource":"apiservices","name":"v1.template.openshift.io"}],"extension":null}}config/clusteroperator/olm.json0000640000000000000000000001420415114274061015426 0ustar0000000000000000{"kind":"ClusterOperator","apiVersion":"config.openshift.io/v1","metadata":{"name":"olm","uid":"183e2b59-e95f-476e-8c63-d724d3987399","resourceVersion":"7299","generation":1,"creationTimestamp":"2025-12-04T11:31:13Z","annotations":{"capability.openshift.io/name":"OperatorLifecycleManagerV1","include.release.openshift.io/self-managed-high-availability":"true"},"ownerReferences":[{"apiVersion":"config.openshift.io/v1","kind":"ClusterVersion","name":"version","uid":"a151bf9a-21b6-4d33-89ec-0e4f83e8eab6","controller":true}]},"spec":{},"status":{"conditions":[{"type":"Degraded","status":"False","lastTransitionTime":"2025-12-04T11:37:58Z","reason":"AsExpected","message":"All is well"},{"type":"Progressing","status":"False","lastTransitionTime":"2025-12-04T11:38:34Z","reason":"AsExpected","message":"All is well"},{"type":"Available","status":"True","lastTransitionTime":"2025-12-04T11:38:34Z","reason":"AsExpected","message":"All is well"},{"type":"Upgradeable","status":"True","lastTransitionTime":"2025-12-04T11:37:58Z","reason":"AsExpected","message":"All is well"},{"type":"EvaluationConditionsDetected","status":"Unknown","lastTransitionTime":"2025-12-04T11:37:58Z","reason":"NoData"}],"versions":[{"name":"operator","version":"4.18.29"}],"relatedObjects":[{"group":"","resource":"namespaces","name":"openshift-catalogd"},{"group":"apiextensions.k8s.io","resource":"customresourcedefinitions","name":"clustercatalogs.olm.operatorframework.io"},{"group":"","resource":"serviceaccounts","namespace":"openshift-catalogd","name":"catalogd-controller-manager"},{"group":"rbac.authorization.k8s.io","resource":"roles","namespace":"openshift-catalogd","name":"catalogd-leader-election-role"},{"group":"rbac.authorization.k8s.io","resource":"roles","namespace":"openshift-config","name":"catalogd-manager-role"},{"group":"rbac.authorization.k8s.io","resource":"clusterroles","name":"catalogd-manager-role"},{"group":"rbac.authorization.k8s.io","resource":"clusterroles","name":"catalogd-metrics-reader"},{"group":"rbac.authorization.k8s.io","resource":"clusterroles","name":"catalogd-proxy-role"},{"group":"rbac.authorization.k8s.io","resource":"rolebindings","namespace":"openshift-catalogd","name":"catalogd-leader-election-rolebinding"},{"group":"rbac.authorization.k8s.io","resource":"rolebindings","namespace":"openshift-config","name":"catalogd-manager-rolebinding"},{"group":"rbac.authorization.k8s.io","resource":"clusterrolebindings","name":"catalogd-manager-rolebinding"},{"group":"rbac.authorization.k8s.io","resource":"clusterrolebindings","name":"catalogd-proxy-rolebinding"},{"group":"","resource":"configmaps","namespace":"openshift-catalogd","name":"catalogd-trusted-ca-bundle"},{"group":"","resource":"services","namespace":"openshift-catalogd","name":"catalogd-service"},{"group":"apps","resource":"deployments","namespace":"openshift-catalogd","name":"catalogd-controller-manager"},{"group":"olm.operatorframework.io","resource":"clustercatalogs","name":"openshift-certified-operators"},{"group":"olm.operatorframework.io","resource":"clustercatalogs","name":"openshift-community-operators"},{"group":"olm.operatorframework.io","resource":"clustercatalogs","name":"openshift-redhat-marketplace"},{"group":"olm.operatorframework.io","resource":"clustercatalogs","name":"openshift-redhat-operators"},{"group":"admissionregistration.k8s.io","resource":"mutatingwebhookconfigurations","name":"catalogd-mutating-webhook-configuration"},{"group":"","resource":"namespaces","name":"openshift-operator-controller"},{"group":"apiextensions.k8s.io","resource":"customresourcedefinitions","name":"clusterextensions.olm.operatorframework.io"},{"group":"","resource":"serviceaccounts","namespace":"openshift-operator-controller","name":"operator-controller-controller-manager"},{"group":"rbac.authorization.k8s.io","resource":"roles","namespace":"openshift-config","name":"operator-controller-manager-role"},{"group":"rbac.authorization.k8s.io","resource":"roles","namespace":"openshift-operator-controller","name":"operator-controller-leader-election-role"},{"group":"rbac.authorization.k8s.io","resource":"roles","namespace":"openshift-operator-controller","name":"operator-controller-manager-role"},{"group":"rbac.authorization.k8s.io","resource":"clusterroles","name":"operator-controller-clusterextension-editor-role"},{"group":"rbac.authorization.k8s.io","resource":"clusterroles","name":"operator-controller-clusterextension-viewer-role"},{"group":"rbac.authorization.k8s.io","resource":"clusterroles","name":"operator-controller-extension-editor-role"},{"group":"rbac.authorization.k8s.io","resource":"clusterroles","name":"operator-controller-extension-viewer-role"},{"group":"rbac.authorization.k8s.io","resource":"clusterroles","name":"operator-controller-manager-role"},{"group":"rbac.authorization.k8s.io","resource":"clusterroles","name":"operator-controller-metrics-reader"},{"group":"rbac.authorization.k8s.io","resource":"clusterroles","name":"operator-controller-proxy-role"},{"group":"rbac.authorization.k8s.io","resource":"rolebindings","namespace":"openshift-config","name":"operator-controller-manager-rolebinding"},{"group":"rbac.authorization.k8s.io","resource":"rolebindings","namespace":"openshift-operator-controller","name":"operator-controller-leader-election-rolebinding"},{"group":"rbac.authorization.k8s.io","resource":"rolebindings","namespace":"openshift-operator-controller","name":"operator-controller-manager-rolebinding"},{"group":"rbac.authorization.k8s.io","resource":"clusterrolebindings","name":"operator-controller-manager-rolebinding"},{"group":"rbac.authorization.k8s.io","resource":"clusterrolebindings","name":"operator-controller-proxy-rolebinding"},{"group":"","resource":"configmaps","namespace":"openshift-operator-controller","name":"operator-controller-trusted-ca-bundle"},{"group":"","resource":"services","namespace":"openshift-operator-controller","name":"operator-controller-controller-manager-metrics-service"},{"group":"apps","resource":"deployments","namespace":"openshift-operator-controller","name":"operator-controller-controller-manager"},{"group":"operator.openshift.io","resource":"olms","name":"cluster"},{"group":"","resource":"namespaces","name":"openshift-cluster-olm-operator"}],"extension":null}}config/clusteroperator/node-tuning.json0000640000000000000000000000342115114274061017065 0ustar0000000000000000{"kind":"ClusterOperator","apiVersion":"config.openshift.io/v1","metadata":{"name":"node-tuning","uid":"170838d3-4840-48d2-9c83-5cefce73b2d8","resourceVersion":"5820","generation":1,"creationTimestamp":"2025-12-04T11:31:14Z","annotations":{"capability.openshift.io/name":"NodeTuning","include.release.openshift.io/hypershift":"true","include.release.openshift.io/ibm-cloud-managed":"true","include.release.openshift.io/self-managed-high-availability":"true","include.release.openshift.io/single-node-developer":"true"},"ownerReferences":[{"apiVersion":"config.openshift.io/v1","kind":"ClusterVersion","name":"version","uid":"a151bf9a-21b6-4d33-89ec-0e4f83e8eab6","controller":true}]},"spec":{},"status":{"conditions":[{"type":"Available","status":"True","lastTransitionTime":"2025-12-04T11:38:03Z","reason":"AsExpected","message":"Cluster has deployed 1/1 \"4.18.29\" operands"},{"type":"Progressing","status":"False","lastTransitionTime":"2025-12-04T11:38:03Z","reason":"AsExpected","message":"Cluster version is \"4.18.29\""},{"type":"Degraded","status":"False","lastTransitionTime":"2025-12-04T11:38:01Z","reason":"AsExpected","message":"DaemonSet \"tuned\" available"}],"versions":[{"name":"ocp-tuned","version":"4.18.29"},{"name":"operator","version":"4.18.29"}],"relatedObjects":[{"group":"","resource":"namespaces","name":"openshift-cluster-node-tuning-operator"},{"group":"tuned.openshift.io","resource":"profiles","namespace":"openshift-cluster-node-tuning-operator","name":""},{"group":"tuned.openshift.io","resource":"tuneds","namespace":"openshift-cluster-node-tuning-operator","name":""},{"group":"apps","resource":"daemonsets","namespace":"openshift-cluster-node-tuning-operator","name":"tuned"},{"group":"performance.openshift.io","resource":"performanceprofiles","name":""}],"extension":null}}config/clusteroperator/network.json0000640000000000000000000003737415114274061016345 0ustar0000000000000000{"kind":"ClusterOperator","apiVersion":"config.openshift.io/v1","metadata":{"name":"network","uid":"8293c62c-5754-4cd7-8082-451e35ce16fd","resourceVersion":"14303","generation":1,"creationTimestamp":"2025-12-04T11:31:13Z","annotations":{"include.release.openshift.io/ibm-cloud-managed":"true","include.release.openshift.io/self-managed-high-availability":"true","include.release.openshift.io/single-node-developer":"true","network.operator.openshift.io/last-seen-state":"{\"DaemonsetStates\":[],\"DeploymentStates\":[],\"StatefulsetStates\":[]}","network.operator.openshift.io/relatedClusterObjects":""},"ownerReferences":[{"apiVersion":"config.openshift.io/v1","kind":"ClusterVersion","name":"version","uid":"a151bf9a-21b6-4d33-89ec-0e4f83e8eab6","controller":true}]},"spec":{},"status":{"conditions":[{"type":"ManagementStateDegraded","status":"False","lastTransitionTime":"2025-12-04T11:36:33Z"},{"type":"Degraded","status":"False","lastTransitionTime":"2025-12-04T11:36:19Z"},{"type":"Upgradeable","status":"True","lastTransitionTime":"2025-12-04T11:36:19Z"},{"type":"Progressing","status":"True","lastTransitionTime":"2025-12-04T11:50:17Z","reason":"Deploying","message":"Deployment \"/openshift-multus/multus-admission-controller\" update is being processed (generation 3, observed generation 1)"},{"type":"Available","status":"True","lastTransitionTime":"2025-12-04T11:37:49Z"}],"versions":[{"name":"operator","version":"4.18.29"}],"relatedObjects":[{"group":"","resource":"configmaps","namespace":"openshift-network-operator","name":"applied-cluster"},{"group":"apiextensions.k8s.io","resource":"customresourcedefinitions","name":"network-attachment-definitions.k8s.cni.cncf.io"},{"group":"apiextensions.k8s.io","resource":"customresourcedefinitions","name":"ippools.whereabouts.cni.cncf.io"},{"group":"apiextensions.k8s.io","resource":"customresourcedefinitions","name":"overlappingrangeipreservations.whereabouts.cni.cncf.io"},{"group":"apiextensions.k8s.io","resource":"customresourcedefinitions","name":"nodeslicepools.whereabouts.cni.cncf.io"},{"group":"","resource":"namespaces","name":"openshift-multus"},{"group":"rbac.authorization.k8s.io","resource":"clusterroles","name":"multus"},{"group":"rbac.authorization.k8s.io","resource":"clusterroles","name":"multus-ancillary-tools"},{"group":"","resource":"serviceaccounts","namespace":"openshift-multus","name":"multus"},{"group":"rbac.authorization.k8s.io","resource":"clusterrolebindings","name":"multus-transient"},{"group":"rbac.authorization.k8s.io","resource":"clusterrolebindings","name":"multus-group"},{"group":"","resource":"serviceaccounts","namespace":"openshift-multus","name":"multus-ancillary-tools"},{"group":"rbac.authorization.k8s.io","resource":"clusterrolebindings","name":"multus-ancillary-tools"},{"group":"rbac.authorization.k8s.io","resource":"clusterrolebindings","name":"multus-cluster-readers"},{"group":"rbac.authorization.k8s.io","resource":"clusterrolebindings","name":"multus-whereabouts"},{"group":"rbac.authorization.k8s.io","resource":"rolebindings","namespace":"openshift-multus","name":"multus-whereabouts"},{"group":"rbac.authorization.k8s.io","resource":"clusterroles","name":"whereabouts-cni"},{"group":"rbac.authorization.k8s.io","resource":"roles","namespace":"openshift-multus","name":"whereabouts-cni"},{"group":"rbac.authorization.k8s.io","resource":"clusterroles","name":"net-attach-def-project"},{"group":"","resource":"configmaps","namespace":"openshift-multus","name":"default-cni-sysctl-allowlist"},{"group":"","resource":"configmaps","namespace":"openshift-multus","name":"cni-copy-resources"},{"group":"","resource":"configmaps","namespace":"openshift-multus","name":"whereabouts-config"},{"group":"","resource":"configmaps","namespace":"openshift-multus","name":"multus-daemon-config"},{"group":"apps","resource":"daemonsets","namespace":"openshift-multus","name":"multus"},{"group":"apps","resource":"daemonsets","namespace":"openshift-multus","name":"multus-additional-cni-plugins"},{"group":"","resource":"serviceaccounts","namespace":"openshift-multus","name":"metrics-daemon-sa"},{"group":"rbac.authorization.k8s.io","resource":"clusterroles","name":"metrics-daemon-role"},{"group":"rbac.authorization.k8s.io","resource":"clusterrolebindings","name":"metrics-daemon-sa-rolebinding"},{"group":"apps","resource":"daemonsets","namespace":"openshift-multus","name":"network-metrics-daemon"},{"group":"monitoring.coreos.com","resource":"servicemonitors","namespace":"openshift-multus","name":"monitor-network"},{"group":"","resource":"services","namespace":"openshift-multus","name":"network-metrics-service"},{"group":"rbac.authorization.k8s.io","resource":"roles","namespace":"openshift-multus","name":"prometheus-k8s"},{"group":"rbac.authorization.k8s.io","resource":"rolebindings","namespace":"openshift-multus","name":"prometheus-k8s"},{"group":"","resource":"services","namespace":"openshift-multus","name":"multus-admission-controller"},{"group":"","resource":"serviceaccounts","namespace":"openshift-multus","name":"multus-ac"},{"group":"rbac.authorization.k8s.io","resource":"clusterroles","name":"multus-admission-controller-webhook"},{"group":"rbac.authorization.k8s.io","resource":"clusterrolebindings","name":"multus-admission-controller-webhook"},{"group":"admissionregistration.k8s.io","resource":"validatingwebhookconfigurations","name":"multus.openshift.io"},{"group":"apps","resource":"deployments","namespace":"openshift-multus","name":"multus-admission-controller"},{"group":"monitoring.coreos.com","resource":"servicemonitors","namespace":"openshift-multus","name":"monitor-multus-admission-controller"},{"group":"rbac.authorization.k8s.io","resource":"roles","namespace":"openshift-multus","name":"prometheus-k8s"},{"group":"rbac.authorization.k8s.io","resource":"rolebindings","namespace":"openshift-multus","name":"prometheus-k8s"},{"group":"monitoring.coreos.com","resource":"prometheusrules","namespace":"openshift-multus","name":"prometheus-k8s-rules"},{"group":"","resource":"namespaces","name":"openshift-ovn-kubernetes"},{"group":"apiextensions.k8s.io","resource":"customresourcedefinitions","name":"egressfirewalls.k8s.ovn.org"},{"group":"apiextensions.k8s.io","resource":"customresourcedefinitions","name":"egressips.k8s.ovn.org"},{"group":"apiextensions.k8s.io","resource":"customresourcedefinitions","name":"egressqoses.k8s.ovn.org"},{"group":"apiextensions.k8s.io","resource":"customresourcedefinitions","name":"adminpolicybasedexternalroutes.k8s.ovn.org"},{"group":"apiextensions.k8s.io","resource":"customresourcedefinitions","name":"egressservices.k8s.ovn.org"},{"group":"apiextensions.k8s.io","resource":"customresourcedefinitions","name":"adminnetworkpolicies.policy.networking.k8s.io"},{"group":"apiextensions.k8s.io","resource":"customresourcedefinitions","name":"baselineadminnetworkpolicies.policy.networking.k8s.io"},{"group":"apiextensions.k8s.io","resource":"customresourcedefinitions","name":"ipamclaims.k8s.cni.cncf.io"},{"group":"apiextensions.k8s.io","resource":"customresourcedefinitions","name":"userdefinednetworks.k8s.ovn.org"},{"group":"apiextensions.k8s.io","resource":"customresourcedefinitions","name":"clusteruserdefinednetworks.k8s.ovn.org"},{"group":"","resource":"serviceaccounts","namespace":"openshift-ovn-kubernetes","name":"ovn-kubernetes-node"},{"group":"rbac.authorization.k8s.io","resource":"roles","namespace":"openshift-ovn-kubernetes","name":"openshift-ovn-kubernetes-node-limited"},{"group":"rbac.authorization.k8s.io","resource":"rolebindings","namespace":"openshift-ovn-kubernetes","name":"openshift-ovn-kubernetes-nodes-identity-limited"},{"group":"rbac.authorization.k8s.io","resource":"clusterroles","name":"openshift-ovn-kubernetes-node-limited"},{"group":"rbac.authorization.k8s.io","resource":"clusterrolebindings","name":"openshift-ovn-kubernetes-node-identity-limited"},{"group":"rbac.authorization.k8s.io","resource":"clusterroles","name":"openshift-ovn-kubernetes-kube-rbac-proxy"},{"group":"rbac.authorization.k8s.io","resource":"clusterrolebindings","name":"openshift-ovn-kubernetes-node-kube-rbac-proxy"},{"group":"","resource":"configmaps","namespace":"openshift-ovn-kubernetes","name":"ovnkube-config"},{"group":"","resource":"serviceaccounts","namespace":"openshift-ovn-kubernetes","name":"ovn-kubernetes-control-plane"},{"group":"rbac.authorization.k8s.io","resource":"clusterroles","name":"openshift-ovn-kubernetes-control-plane-limited"},{"group":"rbac.authorization.k8s.io","resource":"clusterrolebindings","name":"openshift-ovn-kubernetes-control-plane-limited"},{"group":"rbac.authorization.k8s.io","resource":"roles","namespace":"openshift-ovn-kubernetes","name":"openshift-ovn-kubernetes-control-plane-limited"},{"group":"rbac.authorization.k8s.io","resource":"rolebindings","namespace":"openshift-ovn-kubernetes","name":"openshift-ovn-kubernetes-control-plane-limited"},{"group":"network.operator.openshift.io","resource":"operatorpkis","namespace":"openshift-ovn-kubernetes","name":"ovn"},{"group":"network.operator.openshift.io","resource":"operatorpkis","namespace":"openshift-ovn-kubernetes","name":"signer"},{"group":"flowcontrol.apiserver.k8s.io","resource":"flowschemas","name":"openshift-ovn-kubernetes"},{"group":"rbac.authorization.k8s.io","resource":"clusterroles","name":"openshift-ovn-kubernetes-cluster-reader"},{"group":"","resource":"configmaps","namespace":"openshift-ovn-kubernetes","name":"ovnkube-script-lib"},{"group":"rbac.authorization.k8s.io","resource":"clusterroles","name":"openshift-ovn-kubernetes-udn-editor"},{"group":"rbac.authorization.k8s.io","resource":"clusterroles","name":"openshift-ovn-kubernetes-udn-viewer"},{"group":"monitoring.coreos.com","resource":"prometheusrules","namespace":"openshift-ovn-kubernetes","name":"master-rules"},{"group":"monitoring.coreos.com","resource":"prometheusrules","namespace":"openshift-ovn-kubernetes","name":"networking-rules"},{"group":"","resource":"configmaps","namespace":"openshift-config-managed","name":"openshift-network-features"},{"group":"monitoring.coreos.com","resource":"servicemonitors","namespace":"openshift-ovn-kubernetes","name":"monitor-ovn-control-plane-metrics"},{"group":"","resource":"services","namespace":"openshift-ovn-kubernetes","name":"ovn-kubernetes-control-plane"},{"group":"monitoring.coreos.com","resource":"servicemonitors","namespace":"openshift-ovn-kubernetes","name":"monitor-ovn-node"},{"group":"","resource":"services","namespace":"openshift-ovn-kubernetes","name":"ovn-kubernetes-node"},{"group":"rbac.authorization.k8s.io","resource":"roles","namespace":"openshift-ovn-kubernetes","name":"prometheus-k8s"},{"group":"rbac.authorization.k8s.io","resource":"rolebindings","namespace":"openshift-ovn-kubernetes","name":"prometheus-k8s"},{"group":"","resource":"namespaces","name":"openshift-host-network"},{"group":"","resource":"resourcequotas","namespace":"openshift-host-network","name":"host-network-namespace-quotas"},{"group":"apps","resource":"deployments","namespace":"openshift-ovn-kubernetes","name":"ovnkube-control-plane"},{"group":"apps","resource":"daemonsets","namespace":"openshift-ovn-kubernetes","name":"ovnkube-node"},{"group":"admissionregistration.k8s.io","resource":"validatingadmissionpolicies","name":"user-defined-networks-namespace-label"},{"group":"admissionregistration.k8s.io","resource":"validatingadmissionpolicybindings","name":"user-defined-networks-namespace-label-binding"},{"group":"","resource":"namespaces","name":"openshift-network-diagnostics"},{"group":"","resource":"serviceaccounts","namespace":"openshift-network-diagnostics","name":"network-diagnostics"},{"group":"rbac.authorization.k8s.io","resource":"roles","namespace":"openshift-network-diagnostics","name":"network-diagnostics"},{"group":"rbac.authorization.k8s.io","resource":"rolebindings","namespace":"openshift-network-diagnostics","name":"network-diagnostics"},{"group":"rbac.authorization.k8s.io","resource":"clusterroles","name":"network-diagnostics"},{"group":"rbac.authorization.k8s.io","resource":"clusterrolebindings","name":"network-diagnostics"},{"group":"rbac.authorization.k8s.io","resource":"rolebindings","namespace":"kube-system","name":"network-diagnostics"},{"group":"apps","resource":"deployments","namespace":"openshift-network-diagnostics","name":"network-check-source"},{"group":"","resource":"services","namespace":"openshift-network-diagnostics","name":"network-check-source"},{"group":"monitoring.coreos.com","resource":"servicemonitors","namespace":"openshift-network-diagnostics","name":"network-check-source"},{"group":"rbac.authorization.k8s.io","resource":"roles","namespace":"openshift-network-diagnostics","name":"prometheus-k8s"},{"group":"rbac.authorization.k8s.io","resource":"rolebindings","namespace":"openshift-network-diagnostics","name":"prometheus-k8s"},{"group":"apps","resource":"daemonsets","namespace":"openshift-network-diagnostics","name":"network-check-target"},{"group":"","resource":"services","namespace":"openshift-network-diagnostics","name":"network-check-target"},{"group":"rbac.authorization.k8s.io","resource":"roles","namespace":"openshift-config-managed","name":"openshift-network-public-role"},{"group":"rbac.authorization.k8s.io","resource":"rolebindings","namespace":"openshift-config-managed","name":"openshift-network-public-role-binding"},{"group":"","resource":"namespaces","name":"openshift-network-node-identity"},{"group":"","resource":"serviceaccounts","namespace":"openshift-network-node-identity","name":"network-node-identity"},{"group":"rbac.authorization.k8s.io","resource":"clusterrolebindings","name":"network-node-identity"},{"group":"rbac.authorization.k8s.io","resource":"clusterroles","name":"network-node-identity"},{"group":"rbac.authorization.k8s.io","resource":"rolebindings","namespace":"openshift-network-node-identity","name":"network-node-identity-leases"},{"group":"rbac.authorization.k8s.io","resource":"roles","namespace":"openshift-network-node-identity","name":"network-node-identity-leases"},{"group":"rbac.authorization.k8s.io","resource":"rolebindings","namespace":"openshift-network-node-identity","name":"system:openshift:scc:hostnetwork-v2"},{"group":"","resource":"configmaps","namespace":"openshift-network-node-identity","name":"ovnkube-identity-cm"},{"group":"network.operator.openshift.io","resource":"operatorpkis","namespace":"openshift-network-node-identity","name":"network-node-identity"},{"group":"admissionregistration.k8s.io","resource":"validatingwebhookconfigurations","name":"network-node-identity.openshift.io"},{"group":"apps","resource":"daemonsets","namespace":"openshift-network-node-identity","name":"network-node-identity"},{"group":"monitoring.coreos.com","resource":"prometheusrules","namespace":"openshift-network-operator","name":"openshift-network-operator-ipsec-rules"},{"group":"rbac.authorization.k8s.io","resource":"clusterroles","name":"openshift-iptables-alerter"},{"group":"","resource":"serviceaccounts","namespace":"openshift-network-operator","name":"iptables-alerter"},{"group":"rbac.authorization.k8s.io","resource":"clusterrolebindings","name":"openshift-iptables-alerter"},{"group":"","resource":"configmaps","namespace":"openshift-network-operator","name":"iptables-alerter-script"},{"group":"apps","resource":"daemonsets","namespace":"openshift-network-operator","name":"iptables-alerter"},{"group":"","resource":"namespaces","name":"openshift-network-console"},{"group":"","resource":"configmaps","namespace":"openshift-network-console","name":"networking-console-plugin"},{"group":"apps","resource":"deployments","namespace":"openshift-network-console","name":"networking-console-plugin"},{"group":"","resource":"services","namespace":"openshift-network-console","name":"networking-console-plugin"},{"group":"console.openshift.io","resource":"consoleplugins","name":"networking-console-plugin"},{"group":"","resource":"namespaces","name":"openshift-network-operator"},{"group":"operator.openshift.io","resource":"networks","name":"cluster"},{"group":"","resource":"namespaces","name":"openshift-cloud-network-config-controller"}],"extension":null}}config/clusteroperator/monitoring.json0000640000000000000000000000457115114274061017032 0ustar0000000000000000{"kind":"ClusterOperator","apiVersion":"config.openshift.io/v1","metadata":{"name":"monitoring","uid":"c62f7769-8dd3-460e-a12a-4ebf9db8d775","resourceVersion":"14231","generation":1,"creationTimestamp":"2025-12-04T11:31:13Z","annotations":{"include.release.openshift.io/hypershift":"true","include.release.openshift.io/ibm-cloud-managed":"true","include.release.openshift.io/self-managed-high-availability":"true","include.release.openshift.io/single-node-developer":"true"},"ownerReferences":[{"apiVersion":"config.openshift.io/v1","kind":"ClusterVersion","name":"version","uid":"a151bf9a-21b6-4d33-89ec-0e4f83e8eab6","controller":true}]},"spec":{},"status":{"conditions":[{"type":"Available","status":"False","lastTransitionTime":"2025-12-04T12:01:16Z","reason":"UpdatingPrometheusOperatorFailed","message":"UpdatingPrometheusOperator: reconciling Prometheus Operator Admission Webhook Deployment failed: updating Deployment object failed: waiting for DeploymentRollout of openshift-monitoring/prometheus-operator-admission-webhook: context deadline exceeded: got 1 unavailable replicas"},{"type":"Degraded","status":"True","lastTransitionTime":"2025-12-04T12:01:16Z","reason":"UpdatingPrometheusOperatorFailed","message":"UpdatingPrometheusOperator: reconciling Prometheus Operator Admission Webhook Deployment failed: updating Deployment object failed: waiting for DeploymentRollout of openshift-monitoring/prometheus-operator-admission-webhook: context deadline exceeded: got 1 unavailable replicas"},{"type":"Progressing","status":"True","lastTransitionTime":"2025-12-04T11:38:33Z","reason":"RollOutInProgress","message":"Rolling out the stack."},{"type":"Upgradeable","status":"True","lastTransitionTime":"2025-12-04T11:38:33Z"}],"relatedObjects":[{"group":"","resource":"namespaces","name":"openshift-monitoring"},{"group":"","resource":"namespaces","name":"openshift-user-workload-monitoring"},{"group":"monitoring.coreos.com","resource":"servicemonitors","name":""},{"group":"monitoring.coreos.com","resource":"podmonitors","name":""},{"group":"monitoring.coreos.com","resource":"prometheusrules","name":""},{"group":"monitoring.coreos.com","resource":"alertmanagers","name":""},{"group":"monitoring.coreos.com","resource":"prometheuses","name":""},{"group":"monitoring.coreos.com","resource":"thanosrulers","name":""},{"group":"monitoring.coreos.com","resource":"alertmanagerconfigs","name":""}],"extension":null}}config/clusteroperator/marketplace.json0000640000000000000000000000313315114274061017126 0ustar0000000000000000{"kind":"ClusterOperator","apiVersion":"config.openshift.io/v1","metadata":{"name":"marketplace","uid":"3d5df5a2-faf3-4dba-9802-66079ee67b51","resourceVersion":"7178","generation":1,"creationTimestamp":"2025-12-04T11:31:13Z","annotations":{"capability.openshift.io/name":"marketplace","include.release.openshift.io/hypershift":"true","include.release.openshift.io/ibm-cloud-managed":"true","include.release.openshift.io/self-managed-high-availability":"true","include.release.openshift.io/single-node-developer":"true"},"ownerReferences":[{"apiVersion":"config.openshift.io/v1","kind":"ClusterVersion","name":"version","uid":"a151bf9a-21b6-4d33-89ec-0e4f83e8eab6","controller":true}]},"spec":{},"status":{"conditions":[{"type":"Progressing","status":"False","lastTransitionTime":"2025-12-04T11:38:33Z","reason":"OperatorAvailable","message":"Successfully progressed to release version: 4.18.29"},{"type":"Available","status":"True","lastTransitionTime":"2025-12-04T11:38:33Z","reason":"OperatorAvailable","message":"Available release version: 4.18.29"},{"type":"Upgradeable","status":"True","lastTransitionTime":"2025-12-04T11:38:33Z","reason":"OperatorAvailable","message":"Marketplace is upgradeable"},{"type":"Degraded","status":"False","lastTransitionTime":"2025-12-04T11:38:33Z","reason":"OperatorAvailable","message":"Available release version: 4.18.29"}],"versions":[{"name":"operator","version":"4.18.29"}],"relatedObjects":[{"group":"","resource":"namespaces","name":"openshift-marketplace"},{"group":"operators.coreos.com","resource":"catalogsources","namespace":"openshift-marketplace","name":""}],"extension":null}}config/clusteroperator/machine-config.json0000640000000000000000000000507115114274061017510 0ustar0000000000000000{"kind":"ClusterOperator","apiVersion":"config.openshift.io/v1","metadata":{"name":"machine-config","uid":"cd635ec9-c3e7-4fed-aa22-37ef2b395ab1","resourceVersion":"10309","generation":1,"creationTimestamp":"2025-12-04T11:31:14Z","annotations":{"exclude.release.openshift.io/internal-openshift-hosted":"true","include.release.openshift.io/self-managed-high-availability":"true","include.release.openshift.io/single-node-developer":"true"},"ownerReferences":[{"apiVersion":"config.openshift.io/v1","kind":"ClusterVersion","name":"version","uid":"a151bf9a-21b6-4d33-89ec-0e4f83e8eab6","controller":true}]},"spec":{},"status":{"conditions":[{"type":"Progressing","status":"True","lastTransitionTime":"2025-12-04T11:45:11Z","message":"Unable to apply 4.18.29"},{"type":"Degraded","status":"True","lastTransitionTime":"2025-12-04T11:45:24Z","reason":"MachineConfigDaemonFailed","message":"Unable to apply 4.18.29: failed to apply machine config daemon manifests: Internal error occurred: admission plugin \"authorization.openshift.io/RestrictSubjectBindings\" failed to complete validation in 13s"},{"type":"Available","status":"True","lastTransitionTime":"2025-12-04T11:45:24Z","reason":"AsExpected","message":"Cluster has deployed []"},{"type":"Upgradeable","status":"True","lastTransitionTime":"2025-12-04T11:45:24Z","reason":"AsExpected"},{"type":"EvaluationConditionsDetected","status":"False","lastTransitionTime":"2025-12-04T11:45:24Z","reason":"AsExpected"}],"relatedObjects":[{"group":"","resource":"namespaces","name":"openshift-machine-config-operator"},{"group":"machineconfiguration.openshift.io","resource":"machineconfigpools","name":""},{"group":"machineconfiguration.openshift.io","resource":"controllerconfigs","name":""},{"group":"machineconfiguration.openshift.io","resource":"kubeletconfigs","name":""},{"group":"machineconfiguration.openshift.io","resource":"containerruntimeconfigs","name":""},{"group":"machineconfiguration.openshift.io","resource":"machineconfigs","name":""},{"group":"operator.openshift.io","resource":"machineconfigurations","name":""},{"group":"","resource":"nodes","name":""},{"group":"","resource":"namespaces","name":"openshift-kni-infra"},{"group":"","resource":"namespaces","name":"openshift-openstack-infra"},{"group":"","resource":"namespaces","name":"openshift-ovirt-infra"},{"group":"","resource":"namespaces","name":"openshift-vsphere-infra"},{"group":"","resource":"namespaces","name":"openshift-nutanix-infra"},{"group":"","resource":"namespaces","name":"openshift-cloud-platform-infra"}],"extension":{"master":"\u003cunknown\u003e","worker":"\u003cunknown\u003e"}}}config/clusteroperator/machine-approver.json0000640000000000000000000000253315114274061020101 0ustar0000000000000000{"kind":"ClusterOperator","apiVersion":"config.openshift.io/v1","metadata":{"name":"machine-approver","uid":"355186e7-6d05-4443-a829-656eb426ecb7","resourceVersion":"10318","generation":1,"creationTimestamp":"2025-12-04T11:31:13Z","annotations":{"exclude.release.openshift.io/internal-openshift-hosted":"true","include.release.openshift.io/self-managed-high-availability":"true","include.release.openshift.io/single-node-developer":"true"},"ownerReferences":[{"apiVersion":"config.openshift.io/v1","kind":"ClusterVersion","name":"version","uid":"a151bf9a-21b6-4d33-89ec-0e4f83e8eab6","controller":true}]},"spec":{},"status":{"conditions":[{"type":"Available","status":"True","lastTransitionTime":"2025-12-04T11:45:52Z","reason":"AsExpected","message":"Cluster Machine Approver is available at 4.18.29"},{"type":"Degraded","status":"False","lastTransitionTime":"2025-12-04T11:45:52Z","reason":"AsExpected"},{"type":"Progressing","status":"False","lastTransitionTime":"2025-12-04T11:45:52Z","reason":"AsExpected"},{"type":"Upgradeable","status":"True","lastTransitionTime":"2025-12-04T11:45:52Z","reason":"AsExpected"}],"versions":[{"name":"operator","version":"4.18.29"}],"relatedObjects":[{"group":"","resource":"namespaces","name":"openshift-cluster-machine-approver"},{"group":"certificates.k8s.io","resource":"certificatesigningrequests","name":""}],"extension":null}}config/clusteroperator/machine-api.json0000640000000000000000000000371415114274061017016 0ustar0000000000000000{"kind":"ClusterOperator","apiVersion":"config.openshift.io/v1","metadata":{"name":"machine-api","uid":"5b65b83c-22ce-4106-8f70-345490bde8b8","resourceVersion":"10333","generation":1,"creationTimestamp":"2025-12-04T11:31:13Z","annotations":{"capability.openshift.io/name":"MachineAPI","exclude.release.openshift.io/internal-openshift-hosted":"true","include.release.openshift.io/self-managed-high-availability":"true","include.release.openshift.io/single-node-developer":"true"},"ownerReferences":[{"apiVersion":"config.openshift.io/v1","kind":"ClusterVersion","name":"version","uid":"a151bf9a-21b6-4d33-89ec-0e4f83e8eab6","controller":true}]},"spec":{},"status":{"conditions":[{"type":"Progressing","status":"False","lastTransitionTime":"2025-12-04T11:45:53Z","reason":"AsExpected"},{"type":"Degraded","status":"False","lastTransitionTime":"2025-12-04T11:45:52Z","reason":"AsExpected"},{"type":"Available","status":"True","lastTransitionTime":"2025-12-04T11:45:53Z","reason":"AsExpected","message":"Cluster Machine API Operator is in NoOp mode"},{"type":"Upgradeable","status":"True","lastTransitionTime":"2025-12-04T11:45:53Z"}],"versions":[{"name":"operator","version":"4.18.29"}],"relatedObjects":[{"group":"","resource":"namespaces","name":"openshift-machine-api"},{"group":"machine.openshift.io","resource":"machines","namespace":"openshift-machine-api","name":""},{"group":"machine.openshift.io","resource":"machinesets","namespace":"openshift-machine-api","name":""},{"group":"machine.openshift.io","resource":"machinehealthchecks","namespace":"openshift-machine-api","name":""},{"group":"rbac.authorization.k8s.io","resource":"roles","namespace":"openshift-machine-api","name":""},{"group":"rbac.authorization.k8s.io","resource":"clusterroles","name":"machine-api-operator"},{"group":"rbac.authorization.k8s.io","resource":"clusterroles","name":"machine-api-controllers"},{"group":"metal3.io","resource":"baremetalhosts","namespace":"openshift-machine-api","name":""}],"extension":null}}config/clusteroperator/kube-storage-version-migrator.json0000640000000000000000000000325315114274061022536 0ustar0000000000000000{"kind":"ClusterOperator","apiVersion":"config.openshift.io/v1","metadata":{"name":"kube-storage-version-migrator","uid":"ec3d55fb-c5d2-4fe0-aa75-1fb01aa20473","resourceVersion":"4901","generation":1,"creationTimestamp":"2025-12-04T11:31:13Z","annotations":{"include.release.openshift.io/hypershift":"true","include.release.openshift.io/ibm-cloud-managed":"true","include.release.openshift.io/self-managed-high-availability":"true"},"ownerReferences":[{"apiVersion":"config.openshift.io/v1","kind":"ClusterVersion","name":"version","uid":"a151bf9a-21b6-4d33-89ec-0e4f83e8eab6","controller":true}]},"spec":{},"status":{"conditions":[{"type":"Degraded","status":"False","lastTransitionTime":"2025-12-04T11:37:45Z","reason":"AsExpected","message":"All is well"},{"type":"Progressing","status":"False","lastTransitionTime":"2025-12-04T11:37:53Z","reason":"AsExpected","message":"All is well"},{"type":"Available","status":"True","lastTransitionTime":"2025-12-04T11:37:53Z","reason":"AsExpected","message":"All is well"},{"type":"Upgradeable","status":"True","lastTransitionTime":"2025-12-04T11:37:44Z","reason":"AsExpected","message":"All is well"},{"type":"EvaluationConditionsDetected","status":"Unknown","lastTransitionTime":"2025-12-04T11:37:44Z","reason":"NoData"}],"versions":[{"name":"operator","version":"4.18.29"}],"relatedObjects":[{"group":"operator.openshift.io","resource":"kubestorageversionmigrators","name":"cluster"},{"group":"migration.k8s.io","resource":"storageversionmigrations","name":""},{"group":"","resource":"namespaces","name":"openshift-kube-storage-version-migrator"},{"group":"","resource":"namespaces","name":"openshift-kube-storage-version-migrator-operator"}],"extension":null}}config/clusteroperator/kube-scheduler.json0000640000000000000000000000414615114274061017545 0ustar0000000000000000{"kind":"ClusterOperator","apiVersion":"config.openshift.io/v1","metadata":{"name":"kube-scheduler","uid":"26ea0a27-fc35-4f21-88ba-acb7adcdf30c","resourceVersion":"12126","generation":1,"creationTimestamp":"2025-12-04T11:31:13Z","annotations":{"exclude.release.openshift.io/internal-openshift-hosted":"true","include.release.openshift.io/self-managed-high-availability":"true","include.release.openshift.io/single-node-developer":"true"},"ownerReferences":[{"apiVersion":"config.openshift.io/v1","kind":"ClusterVersion","name":"version","uid":"a151bf9a-21b6-4d33-89ec-0e4f83e8eab6","controller":true}]},"spec":{},"status":{"conditions":[{"type":"Degraded","status":"False","lastTransitionTime":"2025-12-04T11:37:45Z","reason":"AsExpected","message":"NodeControllerDegraded: All master nodes are ready"},{"type":"Progressing","status":"False","lastTransitionTime":"2025-12-04T11:50:21Z","reason":"AsExpected","message":"NodeInstallerProgressing: 1 node is at revision 4"},{"type":"Available","status":"True","lastTransitionTime":"2025-12-04T11:50:21Z","reason":"AsExpected","message":"StaticPodsAvailable: 1 nodes are active; 1 node is at revision 4"},{"type":"Upgradeable","status":"True","lastTransitionTime":"2025-12-04T11:37:45Z","reason":"AsExpected","message":"All is well"},{"type":"EvaluationConditionsDetected","status":"Unknown","lastTransitionTime":"2025-12-04T11:37:44Z","reason":"NoData"}],"versions":[{"name":"raw-internal","version":"4.18.29"},{"name":"kube-scheduler","version":"1.31.13"},{"name":"operator","version":"4.18.29"}],"relatedObjects":[{"group":"operator.openshift.io","resource":"kubeschedulers","name":"cluster"},{"group":"config.openshift.io","resource":"schedulers","name":""},{"group":"","resource":"namespaces","name":"openshift-config"},{"group":"","resource":"namespaces","name":"openshift-config-managed"},{"group":"","resource":"namespaces","name":"openshift-kube-scheduler"},{"group":"","resource":"namespaces","name":"openshift-kube-scheduler-operator"},{"group":"controlplane.operator.openshift.io","resource":"podnetworkconnectivitychecks","namespace":"openshift-kube-apiserver","name":""}],"extension":null}}config/clusteroperator/kube-controller-manager.json0000640000000000000000000000511515114274061021357 0ustar0000000000000000{"kind":"ClusterOperator","apiVersion":"config.openshift.io/v1","metadata":{"name":"kube-controller-manager","uid":"3c89b53b-1e96-4289-87bd-dc309074b069","resourceVersion":"10913","generation":1,"creationTimestamp":"2025-12-04T11:31:13Z","annotations":{"exclude.release.openshift.io/internal-openshift-hosted":"true","include.release.openshift.io/self-managed-high-availability":"true","include.release.openshift.io/single-node-developer":"true"},"ownerReferences":[{"apiVersion":"config.openshift.io/v1","kind":"ClusterVersion","name":"version","uid":"a151bf9a-21b6-4d33-89ec-0e4f83e8eab6","controller":true}]},"spec":{},"status":{"conditions":[{"type":"Degraded","status":"False","lastTransitionTime":"2025-12-04T11:37:45Z","reason":"AsExpected","message":"StaticPodsDegraded: pod/kube-controller-manager-master-0 container \"cluster-policy-controller\" is waiting: CrashLoopBackOff: back-off 1m20s restarting failed container=cluster-policy-controller pod=kube-controller-manager-master-0_openshift-kube-controller-manager(e6b437c60bb18680f4492b00b294e872)\nNodeControllerDegraded: All master nodes are ready"},{"type":"Progressing","status":"True","lastTransitionTime":"2025-12-04T11:38:24Z","reason":"NodeInstaller","message":"NodeInstallerProgressing: 1 node is at revision 0; 0 nodes have achieved new revision 3"},{"type":"Available","status":"False","lastTransitionTime":"2025-12-04T11:37:46Z","reason":"StaticPods_ZeroNodesActive","message":"StaticPodsAvailable: 0 nodes are active; 1 node is at revision 0; 0 nodes have achieved new revision 3"},{"type":"Upgradeable","status":"True","lastTransitionTime":"2025-12-04T11:37:45Z","reason":"AsExpected","message":"All is well"},{"type":"EvaluationConditionsDetected","status":"Unknown","lastTransitionTime":"2025-12-04T11:37:44Z","reason":"NoData"}],"versions":[{"name":"raw-internal","version":"4.18.29"},{"name":"kube-controller-manager","version":"1.31.13"},{"name":"operator","version":"4.18.29"}],"relatedObjects":[{"group":"operator.openshift.io","resource":"kubecontrollermanagers","name":"cluster"},{"group":"","resource":"namespaces","name":"openshift-config"},{"group":"","resource":"namespaces","name":"openshift-config-managed"},{"group":"","resource":"namespaces","name":"openshift-kube-controller-manager"},{"group":"","resource":"namespaces","name":"openshift-kube-controller-manager-operator"},{"group":"","resource":"namespaces","name":"kube-system"},{"group":"certificates.k8s.io","resource":"certificatesigningrequests","name":""},{"group":"","resource":"nodes","name":""},{"group":"config.openshift.io","resource":"nodes","name":"cluster"}],"extension":null}}config/clusteroperator/kube-apiserver.json0000640000000000000000000000516715114274061017573 0ustar0000000000000000{"kind":"ClusterOperator","apiVersion":"config.openshift.io/v1","metadata":{"name":"kube-apiserver","uid":"50c8ed09-2643-47a8-98f7-07e2486754ac","resourceVersion":"12187","generation":1,"creationTimestamp":"2025-12-04T11:31:13Z","annotations":{"exclude.release.openshift.io/internal-openshift-hosted":"true","include.release.openshift.io/self-managed-high-availability":"true","include.release.openshift.io/single-node-developer":"true"},"ownerReferences":[{"apiVersion":"config.openshift.io/v1","kind":"ClusterVersion","name":"version","uid":"a151bf9a-21b6-4d33-89ec-0e4f83e8eab6","controller":true}]},"spec":{},"status":{"conditions":[{"type":"Degraded","status":"False","lastTransitionTime":"2025-12-04T11:37:38Z","reason":"AsExpected","message":"NodeControllerDegraded: All master nodes are ready"},{"type":"Progressing","status":"False","lastTransitionTime":"2025-12-04T12:00:02Z","reason":"AsExpected","message":"NodeInstallerProgressing: 1 node is at revision 3"},{"type":"Available","status":"True","lastTransitionTime":"2025-12-04T12:00:02Z","reason":"AsExpected","message":"StaticPodsAvailable: 1 nodes are active; 1 node is at revision 3"},{"type":"Upgradeable","status":"True","lastTransitionTime":"2025-12-04T11:37:36Z","reason":"AsExpected","message":"KubeletMinorVersionUpgradeable: Kubelet and API server minor versions are synced."},{"type":"EvaluationConditionsDetected","status":"False","lastTransitionTime":"2025-12-04T11:37:36Z","reason":"AsExpected","message":"All is well"}],"versions":[{"name":"raw-internal","version":"4.18.29"},{"name":"operator","version":"4.18.29"},{"name":"kube-apiserver","version":"1.31.13"}],"relatedObjects":[{"group":"operator.openshift.io","resource":"kubeapiservers","name":"cluster"},{"group":"apiextensions.k8s.io","resource":"customresourcedefinitions","name":""},{"group":"security.openshift.io","resource":"securitycontextconstraints","name":""},{"group":"","resource":"namespaces","name":"openshift-config"},{"group":"","resource":"namespaces","name":"openshift-config-managed"},{"group":"","resource":"namespaces","name":"openshift-kube-apiserver-operator"},{"group":"","resource":"namespaces","name":"openshift-kube-apiserver"},{"group":"admissionregistration.k8s.io","resource":"mutatingwebhookconfigurations","name":""},{"group":"admissionregistration.k8s.io","resource":"validatingwebhookconfigurations","name":""},{"group":"controlplane.operator.openshift.io","resource":"podnetworkconnectivitychecks","namespace":"openshift-kube-apiserver","name":""},{"group":"apiserver.openshift.io","resource":"apirequestcounts","name":""},{"group":"config.openshift.io","resource":"nodes","name":"cluster"}],"extension":null}}config/clusteroperator/insights.json0000640000000000000000000000560315114274061016472 0ustar0000000000000000{"kind":"ClusterOperator","apiVersion":"config.openshift.io/v1","metadata":{"name":"insights","uid":"bbf7f23a-c053-4143-ab3f-58a151f0d0eb","resourceVersion":"11348","generation":1,"creationTimestamp":"2025-12-04T11:31:13Z","annotations":{"capability.openshift.io/name":"Insights","exclude.release.openshift.io/internal-openshift-hosted":"true","include.release.openshift.io/hypershift":"true","include.release.openshift.io/ibm-cloud-managed":"true","include.release.openshift.io/self-managed-high-availability":"true","include.release.openshift.io/single-node-developer":"true"},"ownerReferences":[{"apiVersion":"config.openshift.io/v1","kind":"ClusterVersion","name":"version","uid":"a151bf9a-21b6-4d33-89ec-0e4f83e8eab6","controller":true}]},"spec":{},"status":{"conditions":[{"type":"Available","status":"True","lastTransitionTime":"2025-12-04T11:50:34Z","reason":"AsExpected","message":"Insights works as expected"},{"type":"ClusterTransferAvailable","status":"False","lastTransitionTime":"2025-12-04T11:56:34Z","reason":"NoClusterTransfer","message":"no available cluster transfer"},{"type":"Degraded","status":"False","lastTransitionTime":"2025-12-04T11:50:34Z","reason":"AsExpected","message":"Insights works as expected"},{"type":"Disabled","status":"False","lastTransitionTime":"2025-12-04T11:50:34Z","reason":"AsExpected"},{"type":"Progressing","status":"False","lastTransitionTime":"2025-12-04T11:50:34Z","reason":"AsExpected","message":"Monitoring the cluster"},{"type":"RemoteConfigurationAvailable","status":"True","lastTransitionTime":"2025-12-04T11:50:34Z","reason":"AsExpected"},{"type":"RemoteConfigurationValid","status":"True","lastTransitionTime":"2025-12-04T11:50:34Z","reason":"AsExpected"},{"type":"SCAAvailable","status":"True","lastTransitionTime":"2025-12-04T11:56:34Z","reason":"Updated","message":"SCA certs successfully updated in the etc-pki-entitlement secret"},{"type":"Upgradeable","status":"True","lastTransitionTime":"2025-12-04T11:50:34Z","reason":"InsightsUpgradeable","message":"Insights operator can be upgraded"}],"versions":[{"name":"operator","version":"4.18.29"}],"relatedObjects":[{"group":"","resource":"namespaces","name":"openshift-insights"},{"group":"apps","resource":"deployments","namespace":"openshift-insights","name":"insights-operator"},{"group":"","resource":"secrets","namespace":"openshift-config","name":"pull-secret"},{"group":"","resource":"secrets","namespace":"openshift-config","name":"support"},{"group":"","resource":"serviceaccounts","namespace":"openshift-insights","name":"gather"},{"group":"","resource":"serviceaccounts","namespace":"openshift-insights","name":"operator"},{"group":"","resource":"services","namespace":"openshift-insights","name":"metrics"},{"group":"","resource":"configmaps","namespace":"openshift-insights","name":"service-ca-bundle"},{"group":"operator.openshift.io","resource":"insightsoperators","name":"cluster"}],"extension":{"lastReportTime":null}}}config/clusteroperator/ingress.json0000640000000000000000000000520515114274061016312 0ustar0000000000000000{"kind":"ClusterOperator","apiVersion":"config.openshift.io/v1","metadata":{"name":"ingress","uid":"3dfe5dff-a950-4fa5-9e4f-f37e43fa974e","resourceVersion":"13618","generation":1,"creationTimestamp":"2025-12-04T11:31:13Z","annotations":{"capability.openshift.io/name":"Ingress","include.release.openshift.io/ibm-cloud-managed":"true","include.release.openshift.io/self-managed-high-availability":"true","include.release.openshift.io/single-node-developer":"true"},"ownerReferences":[{"apiVersion":"config.openshift.io/v1","kind":"ClusterVersion","name":"version","uid":"a151bf9a-21b6-4d33-89ec-0e4f83e8eab6","controller":true}]},"spec":{},"status":{"conditions":[{"type":"Available","status":"True","lastTransitionTime":"2025-12-04T11:38:22Z","reason":"IngressAvailable","message":"The \"default\" ingress controller reports Available=True."},{"type":"Progressing","status":"True","lastTransitionTime":"2025-12-04T11:38:20Z","reason":"Reconciling","message":"ingresscontroller \"default\" is progressing: IngressControllerProgressing: One or more status conditions indicate progressing: DeploymentRollingOut=True (DeploymentRollingOut: Waiting for router deployment rollout to finish: 0 of 1 updated replica(s) are available...\n)."},{"type":"Degraded","status":"True","lastTransitionTime":"2025-12-04T12:00:31Z","reason":"IngressDegraded","message":"The \"default\" ingress controller reports Degraded=True: DegradedConditions: One or more other status conditions indicate a degraded state: CanaryChecksSucceeding=Unknown (CanaryRouteNotAdmitted: Canary route is not admitted by the default ingress controller)"},{"type":"Upgradeable","status":"True","lastTransitionTime":"2025-12-04T11:38:20Z","reason":"Upgradeable"},{"type":"EvaluationConditionsDetected","status":"False","lastTransitionTime":"2025-12-04T11:38:20Z","reason":"AsExpected"}],"versions":[{"name":"operator","version":"4.18.29"},{"name":"ingress-controller","version":"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:2b3d313c599852b3543ee5c3a62691bd2d1bbad12c2e1c610cd71a1dec6eea32"},{"name":"canary-server","version":"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:831f30660844091d6154e2674d3a9da6f34271bf8a2c40b56f7416066318742b"}],"relatedObjects":[{"group":"","resource":"namespaces","name":"openshift-ingress-operator"},{"group":"operator.openshift.io","resource":"ingresscontrollers","namespace":"openshift-ingress-operator","name":""},{"group":"ingress.operator.openshift.io","resource":"dnsrecords","namespace":"openshift-ingress-operator","name":""},{"group":"","resource":"namespaces","name":"openshift-ingress"},{"group":"","resource":"namespaces","name":"openshift-ingress-canary"}],"extension":null}}config/clusteroperator/image-registry.json0000640000000000000000000000405015114274061017565 0ustar0000000000000000{"kind":"ClusterOperator","apiVersion":"config.openshift.io/v1","metadata":{"name":"image-registry","uid":"b2c0bed0-bf7e-4ddf-b1e2-19ddb005327f","resourceVersion":"14290","generation":1,"creationTimestamp":"2025-12-04T11:31:13Z","annotations":{"capability.openshift.io/name":"ImageRegistry","include.release.openshift.io/hypershift":"true","include.release.openshift.io/ibm-cloud-managed":"true","include.release.openshift.io/self-managed-high-availability":"true","include.release.openshift.io/single-node-developer":"true"},"ownerReferences":[{"apiVersion":"config.openshift.io/v1","kind":"ClusterVersion","name":"version","uid":"a151bf9a-21b6-4d33-89ec-0e4f83e8eab6","controller":true}]},"spec":{},"status":{"conditions":[{"type":"Available","status":"False","lastTransitionTime":"2025-12-04T12:01:39Z","reason":"NodeCADaemonNoAvailableReplicas::Removed","message":"Available: The registry is removed\nNodeCADaemonAvailable: The daemon set node-ca does not have available replicas\nImagePrunerAvailable: Pruner CronJob has been created"},{"type":"Progressing","status":"True","lastTransitionTime":"2025-12-04T12:01:39Z","reason":"NodeCADaemonProgressing::Removed","message":"Progressing: All registry resources are removed\nNodeCADaemonProgressing: The daemon set node-ca is updating node pods"},{"type":"Degraded","status":"False","lastTransitionTime":"2025-12-04T12:00:01Z","reason":"Removed","message":"Degraded: The registry is removed"}],"versions":[{"name":"operator","version":"4.18.29"}],"relatedObjects":[{"group":"imageregistry.operator.openshift.io","resource":"configs","name":"cluster"},{"group":"imageregistry.operator.openshift.io","resource":"imagepruners","name":"cluster"},{"group":"rbac.authorization.k8s.io","resource":"clusterroles","name":"system:registry"},{"group":"rbac.authorization.k8s.io","resource":"clusterrolebindings","name":"registry-registry-role"},{"group":"rbac.authorization.k8s.io","resource":"clusterrolebindings","name":"openshift-image-registry-pruner"},{"group":"","resource":"namespaces","name":"openshift-image-registry"}],"extension":null}}config/clusteroperator/etcd.json0000640000000000000000000000400215114274061015551 0ustar0000000000000000{"kind":"ClusterOperator","apiVersion":"config.openshift.io/v1","metadata":{"name":"etcd","uid":"d67c5ace-b973-4fe0-8139-39e325423625","resourceVersion":"14229","generation":1,"creationTimestamp":"2025-12-04T11:31:13Z","annotations":{"exclude.release.openshift.io/internal-openshift-hosted":"true","include.release.openshift.io/self-managed-high-availability":"true","include.release.openshift.io/single-node-developer":"true"},"ownerReferences":[{"apiVersion":"config.openshift.io/v1","kind":"ClusterVersion","name":"version","uid":"a151bf9a-21b6-4d33-89ec-0e4f83e8eab6","controller":true}]},"spec":{},"status":{"conditions":[{"type":"Degraded","status":"False","lastTransitionTime":"2025-12-04T11:50:06Z","reason":"AsExpected","message":"NodeControllerDegraded: All master nodes are ready\nEtcdMembersDegraded: No unhealthy members found"},{"type":"Progressing","status":"False","lastTransitionTime":"2025-12-04T12:01:15Z","reason":"AsExpected","message":"NodeInstallerProgressing: 1 node is at revision 2\nEtcdMembersProgressing: No unstarted etcd members found"},{"type":"Available","status":"True","lastTransitionTime":"2025-12-04T11:50:09Z","reason":"AsExpected","message":"StaticPodsAvailable: 1 nodes are active; 1 node is at revision 2\nEtcdMembersAvailable: 1 members are available"},{"type":"Upgradeable","status":"True","lastTransitionTime":"2025-12-04T11:37:45Z","reason":"AsExpected","message":"All is well"},{"type":"EvaluationConditionsDetected","status":"Unknown","lastTransitionTime":"2025-12-04T11:37:44Z","reason":"NoData"}],"versions":[{"name":"raw-internal","version":"4.18.29"},{"name":"etcd","version":"4.18.29"},{"name":"operator","version":"4.18.29"}],"relatedObjects":[{"group":"operator.openshift.io","resource":"etcds","name":"cluster"},{"group":"","resource":"namespaces","name":"openshift-config"},{"group":"","resource":"namespaces","name":"openshift-config-managed"},{"group":"","resource":"namespaces","name":"openshift-etcd-operator"},{"group":"","resource":"namespaces","name":"openshift-etcd"}],"extension":null}}config/clusteroperator/dns.json0000640000000000000000000000401015114274061015415 0ustar0000000000000000{"kind":"ClusterOperator","apiVersion":"config.openshift.io/v1","metadata":{"name":"dns","uid":"4d8293fb-d7de-406c-9854-07131e91b9ec","resourceVersion":"10322","generation":1,"creationTimestamp":"2025-12-04T11:31:14Z","annotations":{"include.release.openshift.io/ibm-cloud-managed":"true","include.release.openshift.io/self-managed-high-availability":"true","include.release.openshift.io/single-node-developer":"true"},"ownerReferences":[{"apiVersion":"config.openshift.io/v1","kind":"ClusterVersion","name":"version","uid":"a151bf9a-21b6-4d33-89ec-0e4f83e8eab6","controller":true}]},"spec":{},"status":{"conditions":[{"type":"Available","status":"False","lastTransitionTime":"2025-12-04T11:38:20Z","reason":"DNSUnavailable","message":"DNS \"default\" is unavailable."},{"type":"Progressing","status":"True","lastTransitionTime":"2025-12-04T11:45:53Z","reason":"DNSReportsProgressingIsTrue","message":"DNS \"default\" reports Progressing=True: \"Have 0 available DNS pods, want 1.\""},{"type":"Degraded","status":"True","lastTransitionTime":"2025-12-04T11:38:21Z","reason":"DNSDegraded","message":"DNS default is degraded"},{"type":"Upgradeable","status":"True","lastTransitionTime":"2025-12-04T11:38:20Z","reason":"DNSUpgradeable","message":"DNS default is upgradeable: DNS Operator can be upgraded"}],"versions":[{"name":"operator","version":"4.18.29"},{"name":"coredns","version":"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:eb928c13a46d3fb45f4a881892d023a92d610a5430be0ffd916aaf8da8e7d297"},{"name":"openshift-cli","version":"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:79f99fd6cce984287932edf0d009660bb488d663081f3d62ec3b23bc8bfbf6c2"},{"name":"kube-rbac-proxy","version":"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69ffd8f8dcceedc2d6eb306cea33f8beabc1be1308cd5f4ee8b9a8e3eab9843"}],"relatedObjects":[{"group":"","resource":"namespaces","name":"openshift-dns-operator"},{"group":"operator.openshift.io","resource":"dnses","name":"default"},{"group":"","resource":"namespaces","name":"openshift-dns"}],"extension":null}}config/clusteroperator/csi-snapshot-controller.json0000640000000000000000000000322015114274061021427 0ustar0000000000000000{"kind":"ClusterOperator","apiVersion":"config.openshift.io/v1","metadata":{"name":"csi-snapshot-controller","uid":"949c008d-8fff-4718-87e5-61d19cda5b58","resourceVersion":"13585","generation":1,"creationTimestamp":"2025-12-04T11:31:13Z","annotations":{"capability.openshift.io/name":"CSISnapshot","include.release.openshift.io/hypershift":"true","include.release.openshift.io/ibm-cloud-managed":"true","include.release.openshift.io/self-managed-high-availability":"true","include.release.openshift.io/single-node-developer":"true"},"ownerReferences":[{"apiVersion":"config.openshift.io/v1","kind":"ClusterVersion","name":"version","uid":"a151bf9a-21b6-4d33-89ec-0e4f83e8eab6","controller":true}]},"spec":{},"status":{"conditions":[{"type":"Degraded","status":"False","lastTransitionTime":"2025-12-04T11:37:44Z","reason":"AsExpected","message":"All is well"},{"type":"Progressing","status":"False","lastTransitionTime":"2025-12-04T11:37:53Z","reason":"AsExpected","message":"All is well"},{"type":"Available","status":"True","lastTransitionTime":"2025-12-04T11:37:53Z","reason":"AsExpected","message":"All is well"},{"type":"Upgradeable","status":"True","lastTransitionTime":"2025-12-04T11:37:44Z","reason":"AsExpected","message":"All is well"},{"type":"EvaluationConditionsDetected","status":"Unknown","lastTransitionTime":"2025-12-04T11:37:44Z","reason":"NoData"}],"versions":[{"name":"csi-snapshot-controller","version":"4.18.29"},{"name":"operator","version":"4.18.29"}],"relatedObjects":[{"group":"","resource":"namespaces","name":"openshift-cluster-storage-operator"},{"group":"operator.openshift.io","resource":"csisnapshotcontrollers","name":"cluster"}],"extension":null}}config/clusteroperator/control-plane-machine-set.json0000640000000000000000000000263515114274061021614 0ustar0000000000000000{"kind":"ClusterOperator","apiVersion":"config.openshift.io/v1","metadata":{"name":"control-plane-machine-set","uid":"88136873-c571-464b-a036-d79e80871b78","resourceVersion":"8125","generation":1,"creationTimestamp":"2025-12-04T11:31:13Z","annotations":{"capability.openshift.io/name":"MachineAPI","exclude.release.openshift.io/internal-openshift-hosted":"true","include.release.openshift.io/self-managed-high-availability":"true"},"ownerReferences":[{"apiVersion":"config.openshift.io/v1","kind":"ClusterVersion","name":"version","uid":"a151bf9a-21b6-4d33-89ec-0e4f83e8eab6","controller":true}]},"spec":{},"status":{"conditions":[{"type":"Available","status":"True","lastTransitionTime":"2025-12-04T11:38:41Z","reason":"AsExpected","message":"cluster operator is available"},{"type":"Progressing","status":"False","lastTransitionTime":"2025-12-04T11:38:41Z","reason":"AsExpected"},{"type":"Degraded","status":"False","lastTransitionTime":"2025-12-04T11:38:41Z","reason":"AsExpected"},{"type":"Upgradeable","status":"True","lastTransitionTime":"2025-12-04T11:38:41Z","reason":"AsExpected","message":"cluster operator is upgradable"}],"versions":[{"name":"operator","version":"4.18.29"}],"relatedObjects":[{"group":"","resource":"namespaces","name":"openshift-machine-api"},{"group":"machine.openshift.io","resource":"controlplanemachinesets","name":""},{"group":"machine.openshift.io","resource":"machines","name":""}],"extension":null}}config/clusteroperator/console.json0000640000000000000000000000131215114274061016275 0ustar0000000000000000{"kind":"ClusterOperator","apiVersion":"config.openshift.io/v1","metadata":{"name":"console","uid":"fa7c1a71-877b-4f28-be90-721af80c00b7","resourceVersion":"724","generation":1,"creationTimestamp":"2025-12-04T11:31:13Z","annotations":{"capability.openshift.io/name":"Console","include.release.openshift.io/hypershift":"true","include.release.openshift.io/ibm-cloud-managed":"true","include.release.openshift.io/self-managed-high-availability":"true","include.release.openshift.io/single-node-developer":"true"},"ownerReferences":[{"apiVersion":"config.openshift.io/v1","kind":"ClusterVersion","name":"version","uid":"a151bf9a-21b6-4d33-89ec-0e4f83e8eab6","controller":true}]},"spec":{},"status":{"extension":null}}config/clusteroperator/config-operator.json0000640000000000000000000000311615114274061017735 0ustar0000000000000000{"kind":"ClusterOperator","apiVersion":"config.openshift.io/v1","metadata":{"name":"config-operator","uid":"7303edc8-7d9b-45a6-9e37-16e359d406aa","resourceVersion":"4922","generation":1,"creationTimestamp":"2025-12-04T11:31:13Z","annotations":{"exclude.release.openshift.io/internal-openshift-hosted":"true","include.release.openshift.io/self-managed-high-availability":"true","include.release.openshift.io/single-node-developer":"true"},"ownerReferences":[{"apiVersion":"config.openshift.io/v1","kind":"ClusterVersion","name":"version","uid":"a151bf9a-21b6-4d33-89ec-0e4f83e8eab6","controller":true}]},"spec":{},"status":{"conditions":[{"type":"Degraded","status":"False","lastTransitionTime":"2025-12-04T11:37:53Z","reason":"AsExpected","message":"All is well"},{"type":"Progressing","status":"False","lastTransitionTime":"2025-12-04T11:37:53Z","reason":"AsExpected","message":"All is well"},{"type":"Available","status":"True","lastTransitionTime":"2025-12-04T11:37:53Z","reason":"AsExpected","message":"All is well"},{"type":"Upgradeable","status":"True","lastTransitionTime":"2025-12-04T11:37:53Z","reason":"AsExpected","message":"All is well"},{"type":"EvaluationConditionsDetected","status":"Unknown","lastTransitionTime":"2025-12-04T11:37:53Z","reason":"NoData"}],"versions":[{"name":"feature-gates","version":"4.18.29"},{"name":"operator","version":"4.18.29"}],"relatedObjects":[{"group":"operator.openshift.io","resource":"configs","name":"cluster"},{"group":"","resource":"namespaces","name":"openshift-config"},{"group":"","resource":"namespaces","name":"openshift-config-operator"}],"extension":null}}config/clusteroperator/cluster-autoscaler.json0000640000000000000000000000266015114274061020463 0ustar0000000000000000{"kind":"ClusterOperator","apiVersion":"config.openshift.io/v1","metadata":{"name":"cluster-autoscaler","uid":"ad1a7878-3d5a-4077-a935-12a433435830","resourceVersion":"10453","generation":1,"creationTimestamp":"2025-12-04T11:31:13Z","annotations":{"capability.openshift.io/name":"MachineAPI","exclude.release.openshift.io/internal-openshift-hosted":"true","include.release.openshift.io/self-managed-high-availability":"true","include.release.openshift.io/single-node-developer":"true"},"ownerReferences":[{"apiVersion":"config.openshift.io/v1","kind":"ClusterVersion","name":"version","uid":"a151bf9a-21b6-4d33-89ec-0e4f83e8eab6","controller":true}]},"spec":{},"status":{"conditions":[{"type":"Available","status":"True","lastTransitionTime":"2025-12-04T11:46:04Z","reason":"AsExpected","message":"at version 4.18.29"},{"type":"Progressing","status":"False","lastTransitionTime":"2025-12-04T11:46:04Z"},{"type":"Degraded","status":"False","lastTransitionTime":"2025-12-04T11:46:04Z"},{"type":"Upgradeable","status":"True","lastTransitionTime":"2025-12-04T11:46:04Z"}],"versions":[{"name":"operator","version":"4.18.29"}],"relatedObjects":[{"group":"autoscaling.openshift.io","resource":"machineautoscalers","namespace":"openshift-machine-api","name":""},{"group":"autoscaling.openshift.io","resource":"clusterautoscalers","namespace":"openshift-machine-api","name":""},{"group":"","resource":"namespaces","name":"openshift-machine-api"}],"extension":null}}config/clusteroperator/cloud-credential.json0000640000000000000000000002022415114274061020054 0ustar0000000000000000{"kind":"ClusterOperator","apiVersion":"config.openshift.io/v1","metadata":{"name":"cloud-credential","uid":"799c81dc-e152-4471-a4ee-60ea2a6684a3","resourceVersion":"11002","generation":1,"creationTimestamp":"2025-12-04T11:31:14Z","annotations":{"capability.openshift.io/name":"CloudCredential","exclude.release.openshift.io/internal-openshift-hosted":"true","include.release.openshift.io/self-managed-high-availability":"true"},"ownerReferences":[{"apiVersion":"config.openshift.io/v1","kind":"ClusterVersion","name":"version","uid":"a151bf9a-21b6-4d33-89ec-0e4f83e8eab6","controller":true}]},"spec":{},"status":{"conditions":[{"type":"Available","status":"True","lastTransitionTime":"2025-12-04T11:45:55Z","reason":"AsExpected","message":"All is well"},{"type":"Degraded","status":"False","lastTransitionTime":"2025-12-04T11:45:55Z","reason":"AsExpected","message":"All is well"},{"type":"Progressing","status":"False","lastTransitionTime":"2025-12-04T11:45:55Z","reason":"AsExpected","message":"All is well"},{"type":"Upgradeable","status":"True","lastTransitionTime":"2025-12-04T11:45:55Z"}],"versions":[{"name":"operator","version":"4.18.29"}],"relatedObjects":[{"group":"cloudcredential.openshift.io","resource":"credentialsrequests","namespace":"openshift-cloud-credential-operator","name":"aws-ebs-csi-driver-operator"},{"group":"cloudcredential.openshift.io","resource":"credentialsrequests","namespace":"openshift-cloud-credential-operator","name":"azure-disk-csi-driver-operator"},{"group":"cloudcredential.openshift.io","resource":"credentialsrequests","namespace":"openshift-cloud-credential-operator","name":"azure-file-csi-driver-operator"},{"group":"cloudcredential.openshift.io","resource":"credentialsrequests","namespace":"openshift-cloud-credential-operator","name":"cloud-credential-operator-gcp-ro-creds"},{"group":"cloudcredential.openshift.io","resource":"credentialsrequests","namespace":"openshift-cloud-credential-operator","name":"cloud-credential-operator-iam-ro"},{"group":"operator.openshift.io","resource":"cloudcredentials","name":"cluster"},{"group":"cloudcredential.openshift.io","resource":"credentialsrequests","namespace":"openshift-cloud-credential-operator","name":"ibm-powervs-block-csi-driver-operator"},{"group":"cloudcredential.openshift.io","resource":"credentialsrequests","namespace":"openshift-cloud-credential-operator","name":"ibm-vpc-block-csi-driver-operator"},{"group":"cloudcredential.openshift.io","resource":"credentialsrequests","namespace":"openshift-cloud-credential-operator","name":"manila-csi-driver-operator"},{"group":"cloudcredential.openshift.io","resource":"credentialsrequests","namespace":"openshift-cloud-credential-operator","name":"openshift-azure-cloud-controller-manager"},{"group":"","resource":"namespaces","name":"openshift-cloud-credential-operator"},{"group":"cloudcredential.openshift.io","resource":"credentialsrequests","namespace":"openshift-cloud-credential-operator","name":"openshift-cloud-network-config-controller-aws"},{"group":"cloudcredential.openshift.io","resource":"credentialsrequests","namespace":"openshift-cloud-credential-operator","name":"openshift-cloud-network-config-controller-azure"},{"group":"cloudcredential.openshift.io","resource":"credentialsrequests","namespace":"openshift-cloud-credential-operator","name":"openshift-cloud-network-config-controller-gcp"},{"group":"cloudcredential.openshift.io","resource":"credentialsrequests","namespace":"openshift-cloud-credential-operator","name":"openshift-cloud-network-config-controller-openstack"},{"group":"cloudcredential.openshift.io","resource":"credentialsrequests","namespace":"openshift-cloud-credential-operator","name":"openshift-cluster-csi-drivers"},{"group":"cloudcredential.openshift.io","resource":"credentialsrequests","namespace":"openshift-cloud-credential-operator","name":"openshift-gcp-ccm"},{"group":"cloudcredential.openshift.io","resource":"credentialsrequests","namespace":"openshift-cloud-credential-operator","name":"openshift-gcp-pd-csi-driver-operator"},{"group":"cloudcredential.openshift.io","resource":"credentialsrequests","namespace":"openshift-cloud-credential-operator","name":"openshift-ibm-cloud-controller-manager"},{"group":"cloudcredential.openshift.io","resource":"credentialsrequests","namespace":"openshift-cloud-credential-operator","name":"openshift-image-registry"},{"group":"cloudcredential.openshift.io","resource":"credentialsrequests","namespace":"openshift-cloud-credential-operator","name":"openshift-image-registry-azure"},{"group":"cloudcredential.openshift.io","resource":"credentialsrequests","namespace":"openshift-cloud-credential-operator","name":"openshift-image-registry-gcs"},{"group":"cloudcredential.openshift.io","resource":"credentialsrequests","namespace":"openshift-cloud-credential-operator","name":"openshift-image-registry-ibmcos"},{"group":"cloudcredential.openshift.io","resource":"credentialsrequests","namespace":"openshift-cloud-credential-operator","name":"openshift-image-registry-ibmcos-powervs"},{"group":"cloudcredential.openshift.io","resource":"credentialsrequests","namespace":"openshift-cloud-credential-operator","name":"openshift-image-registry-openstack"},{"group":"cloudcredential.openshift.io","resource":"credentialsrequests","namespace":"openshift-cloud-credential-operator","name":"openshift-ingress"},{"group":"cloudcredential.openshift.io","resource":"credentialsrequests","namespace":"openshift-cloud-credential-operator","name":"openshift-ingress-azure"},{"group":"cloudcredential.openshift.io","resource":"credentialsrequests","namespace":"openshift-cloud-credential-operator","name":"openshift-ingress-gcp"},{"group":"cloudcredential.openshift.io","resource":"credentialsrequests","namespace":"openshift-cloud-credential-operator","name":"openshift-machine-api-aws"},{"group":"cloudcredential.openshift.io","resource":"credentialsrequests","namespace":"openshift-cloud-credential-operator","name":"openshift-machine-api-azure"},{"group":"cloudcredential.openshift.io","resource":"credentialsrequests","namespace":"openshift-cloud-credential-operator","name":"openshift-machine-api-gcp"},{"group":"cloudcredential.openshift.io","resource":"credentialsrequests","namespace":"openshift-cloud-credential-operator","name":"openshift-machine-api-ibmcloud"},{"group":"cloudcredential.openshift.io","resource":"credentialsrequests","namespace":"openshift-cloud-credential-operator","name":"openshift-machine-api-nutanix"},{"group":"cloudcredential.openshift.io","resource":"credentialsrequests","namespace":"openshift-cloud-credential-operator","name":"openshift-machine-api-openstack"},{"group":"cloudcredential.openshift.io","resource":"credentialsrequests","namespace":"openshift-cloud-credential-operator","name":"openshift-machine-api-powervs"},{"group":"cloudcredential.openshift.io","resource":"credentialsrequests","namespace":"openshift-cloud-credential-operator","name":"openshift-machine-api-vsphere"},{"group":"cloudcredential.openshift.io","resource":"credentialsrequests","namespace":"openshift-cloud-credential-operator","name":"openshift-network"},{"group":"cloudcredential.openshift.io","resource":"credentialsrequests","namespace":"openshift-cloud-credential-operator","name":"openshift-nutanix-cloud-controller-manager"},{"group":"cloudcredential.openshift.io","resource":"credentialsrequests","namespace":"openshift-cloud-credential-operator","name":"openshift-openstack-cloud-controller-manager"},{"group":"cloudcredential.openshift.io","resource":"credentialsrequests","namespace":"openshift-cloud-credential-operator","name":"openshift-powervs-cloud-controller-manager"},{"group":"cloudcredential.openshift.io","resource":"credentialsrequests","namespace":"openshift-cloud-credential-operator","name":"openshift-vmware-vsphere-csi-driver-operator"},{"group":"cloudcredential.openshift.io","resource":"credentialsrequests","namespace":"openshift-cloud-credential-operator","name":"openshift-vsphere-cloud-controller-manager"},{"group":"cloudcredential.openshift.io","resource":"credentialsrequests","namespace":"openshift-cloud-credential-operator","name":"openshift-vsphere-problem-detector"},{"group":"cloudcredential.openshift.io","resource":"credentialsrequests","namespace":"openshift-cloud-credential-operator","name":"ovirt-csi-driver-operator"}],"extension":null}}config/clusteroperator/cloud-controller-manager.json0000640000000000000000000000440215114274061021535 0ustar0000000000000000{"kind":"ClusterOperator","apiVersion":"config.openshift.io/v1","metadata":{"name":"cloud-controller-manager","uid":"a1f9ea72-6c76-406e-91c5-55fdbee06109","resourceVersion":"10330","generation":1,"creationTimestamp":"2025-12-04T11:31:13Z","annotations":{"capability.openshift.io/name":"CloudControllerManager","exclude.release.openshift.io/internal-openshift-hosted":"true","include.release.openshift.io/self-managed-high-availability":"true","include.release.openshift.io/single-node-developer":"true"},"ownerReferences":[{"apiVersion":"config.openshift.io/v1","kind":"ClusterVersion","name":"version","uid":"a151bf9a-21b6-4d33-89ec-0e4f83e8eab6","controller":true}]},"spec":{},"status":{"conditions":[{"type":"CloudConfigControllerAvailable","status":"True","lastTransitionTime":"2025-12-04T11:45:33Z","reason":"AsExpected","message":"Cloud Config Controller works as expected"},{"type":"CloudConfigControllerDegraded","status":"False","lastTransitionTime":"2025-12-04T11:45:33Z","reason":"AsExpected","message":"Cloud Config Controller works as expected"},{"type":"TrustedCABundleControllerControllerAvailable","status":"True","lastTransitionTime":"2025-12-04T11:45:53Z","reason":"AsExpected","message":"Trusted CA Bundle Controller works as expected"},{"type":"TrustedCABundleControllerControllerDegraded","status":"False","lastTransitionTime":"2025-12-04T11:45:53Z","reason":"AsExpected","message":"Trusted CA Bundle Controller works as expected"},{"type":"Available","status":"True","lastTransitionTime":"2025-12-04T11:45:53Z","reason":"AsExpected","message":"Cluster Cloud Controller Manager Operator is available at 4.18.29"},{"type":"Progressing","status":"False","lastTransitionTime":"2025-12-04T11:45:53Z","reason":"AsExpected"},{"type":"Degraded","status":"False","lastTransitionTime":"2025-12-04T11:45:53Z","reason":"AsExpected"},{"type":"Upgradeable","status":"True","lastTransitionTime":"2025-12-04T11:45:53Z","reason":"AsExpected"}],"versions":[{"name":"operator","version":"4.18.29"}],"relatedObjects":[{"group":"","resource":"namespaces","name":"openshift-cloud-controller-manager-operator"},{"group":"config.openshift.io","resource":"clusteroperators","name":"cloud-controller-manager"},{"group":"","resource":"namespaces","name":"openshift-cloud-controller-manager"}],"extension":null}}config/clusteroperator/baremetal.json0000640000000000000000000000427115114274061016576 0ustar0000000000000000{"kind":"ClusterOperator","apiVersion":"config.openshift.io/v1","metadata":{"name":"baremetal","uid":"849f88b9-db12-4b61-b4bf-e5c4bb96147e","resourceVersion":"10332","generation":1,"creationTimestamp":"2025-12-04T11:31:13Z","annotations":{"capability.openshift.io/name":"baremetal","exclude.release.openshift.io/internal-openshift-hosted":"true","include.release.openshift.io/self-managed-high-availability":"true","include.release.openshift.io/single-node-developer":"true"},"ownerReferences":[{"apiVersion":"config.openshift.io/v1","kind":"ClusterVersion","name":"version","uid":"a151bf9a-21b6-4d33-89ec-0e4f83e8eab6","controller":true}]},"spec":{},"status":{"conditions":[{"type":"Progressing","status":"False","lastTransitionTime":"2025-12-04T11:45:53Z","reason":"WaitingForProvisioningCR"},{"type":"Degraded","status":"False","lastTransitionTime":"2025-12-04T11:45:53Z"},{"type":"Available","status":"True","lastTransitionTime":"2025-12-04T11:45:53Z","reason":"WaitingForProvisioningCR","message":"Waiting for Provisioning CR on BareMetal Platform"},{"type":"Upgradeable","status":"True","lastTransitionTime":"2025-12-04T11:45:53Z"},{"type":"Disabled","status":"False","lastTransitionTime":"2025-12-04T11:45:53Z"}],"versions":[{"name":"operator","version":"4.18.29"}],"relatedObjects":[{"group":"","resource":"namespaces","name":"openshift-machine-api"},{"group":"metal3.io","resource":"baremetalhosts","namespace":"openshift-machine-api","name":""},{"group":"metal3.io","resource":"provisioning","name":""},{"group":"metal3.io","resource":"hostfirmwaresettings","namespace":"openshift-machine-api","name":""},{"group":"metal3.io","resource":"firmwareschemas","namespace":"openshift-machine-api","name":""},{"group":"metal3.io","resource":"preprovisioningimages","namespace":"openshift-machine-api","name":""},{"group":"metal3.io","resource":"bmceventsubscriptions","namespace":"openshift-machine-api","name":""},{"group":"metal3.io","resource":"hostfirmwarecomponents","namespace":"openshift-machine-api","name":""},{"group":"metal3.io","resource":"dataimages","namespace":"openshift-machine-api","name":""},{"group":"metal3.io","resource":"hostupdatepolicies","namespace":"openshift-machine-api","name":""}],"extension":null}}config/clusteroperator/authentication.json0000640000000000000000000001211415114274061017654 0ustar0000000000000000{"kind":"ClusterOperator","apiVersion":"config.openshift.io/v1","metadata":{"name":"authentication","uid":"7fe42f52-324e-457a-83ca-7919bda73132","resourceVersion":"13155","generation":1,"creationTimestamp":"2025-12-04T11:31:13Z","annotations":{"exclude.release.openshift.io/internal-openshift-hosted":"true","include.release.openshift.io/self-managed-high-availability":"true","include.release.openshift.io/single-node-developer":"true"},"ownerReferences":[{"apiVersion":"config.openshift.io/v1","kind":"ClusterVersion","name":"version","uid":"a151bf9a-21b6-4d33-89ec-0e4f83e8eab6","controller":true}]},"spec":{},"status":{"conditions":[{"type":"Degraded","status":"True","lastTransitionTime":"2025-12-04T11:56:30Z","reason":"IngressStateEndpoints_MissingSubsets::OAuthClientsController_SyncError::OAuthServerDeployment_PreconditionNotFulfilled::OAuthServerRouteEndpointAccessibleController_SyncError::OAuthServerServiceEndpointAccessibleController_SyncError::OAuthServerServiceEndpointsEndpointAccessibleController_SyncError::OAuthServerWorkload_PreconditionNotFulfilled::WellKnownReadyController_SyncError","message":"IngressStateEndpointsDegraded: No subsets found for the endpoints of oauth-server\nOAuthClientsControllerDegraded: no ingress for host oauth-openshift.apps.sno.openstack.lab in route oauth-openshift in namespace openshift-authentication\nOAuthServerDeploymentDegraded: waiting for the oauth-openshift route to contain an admitted ingress: no admitted ingress for route oauth-openshift in namespace openshift-authentication\nOAuthServerDeploymentDegraded: \nOAuthServerRouteEndpointAccessibleControllerDegraded: route \"openshift-authentication/oauth-openshift\": status does not have a valid host address\nOAuthServerServiceEndpointAccessibleControllerDegraded: Get \"https://172.30.202.245:443/healthz\": dial tcp 172.30.202.245:443: connect: connection refused\nOAuthServerServiceEndpointsEndpointAccessibleControllerDegraded: oauth service endpoints are not ready\nOAuthServerWorkloadDegraded: waiting for the oauth-openshift route to contain an admitted ingress: no admitted ingress for route oauth-openshift in namespace openshift-authentication\nOAuthServerWorkloadDegraded: \nWellKnownReadyControllerDegraded: failed to get oauth metadata from openshift-config-managed/oauth-openshift ConfigMap: configmap \"oauth-openshift\" not found (check authentication operator, it is supposed to create this)"},{"type":"Progressing","status":"True","lastTransitionTime":"2025-12-04T11:59:50Z","reason":"WellKnownReadyController"},{"type":"Available","status":"False","lastTransitionTime":"2025-12-04T11:37:45Z","reason":"OAuthServerDeployment_PreconditionNotFulfilled::OAuthServerRouteEndpointAccessibleController_ResourceNotFound::OAuthServerServiceEndpointAccessibleController_EndpointUnavailable::OAuthServerServiceEndpointsEndpointAccessibleController_ResourceNotFound::ReadyIngressNodes_NoReadyIngressNodes::WellKnown_NotReady","message":"OAuthServerRouteEndpointAccessibleControllerAvailable: failed to retrieve route from cache: route.route.openshift.io \"oauth-openshift\" not found\nOAuthServerServiceEndpointAccessibleControllerAvailable: Get \"https://172.30.202.245:443/healthz\": dial tcp 172.30.202.245:443: connect: connection refused\nOAuthServerServiceEndpointsEndpointAccessibleControllerAvailable: endpoints \"oauth-openshift\" not found\nReadyIngressNodesAvailable: Authentication requires functional ingress which requires at least one schedulable and ready node. Got 0 worker nodes, 1 master nodes, 0 custom target nodes (none are schedulable or ready for ingress pods).\nWellKnownAvailable: The well-known endpoint is not yet available: failed to get oauth metadata from openshift-config-managed/oauth-openshift ConfigMap: configmap \"oauth-openshift\" not found (check authentication operator, it is supposed to create this)"},{"type":"Upgradeable","status":"True","lastTransitionTime":"2025-12-04T11:37:45Z","reason":"AsExpected","message":"All is well"},{"type":"EvaluationConditionsDetected","status":"Unknown","lastTransitionTime":"2025-12-04T11:37:45Z","reason":"NoData"}],"versions":[{"name":"operator","version":"4.18.29"},{"name":"oauth-apiserver","version":"4.18.29"}],"relatedObjects":[{"group":"operator.openshift.io","resource":"authentications","name":"cluster"},{"group":"config.openshift.io","resource":"authentications","name":"cluster"},{"group":"config.openshift.io","resource":"infrastructures","name":"cluster"},{"group":"config.openshift.io","resource":"oauths","name":"cluster"},{"group":"route.openshift.io","resource":"routes","namespace":"openshift-authentication","name":"oauth-openshift"},{"group":"","resource":"services","namespace":"openshift-authentication","name":"oauth-openshift"},{"group":"","resource":"namespaces","name":"openshift-config"},{"group":"","resource":"namespaces","name":"openshift-config-managed"},{"group":"","resource":"namespaces","name":"openshift-authentication"},{"group":"","resource":"namespaces","name":"openshift-authentication-operator"},{"group":"","resource":"namespaces","name":"openshift-ingress"},{"group":"","resource":"namespaces","name":"openshift-oauth-apiserver"}],"extension":null}}config/running_containers.json0000640000000000000000000001011215114274061015301 0ustar0000000000000000{"images":{"0":"sha256:c69ffd8f8dcceedc2d6eb306cea33f8beabc1be1308cd5f4ee8b9a8e3eab9843","1":"sha256:e3e65409fc2b27ad0aaeb500a39e264663d2980821f099b830b551785ce4ce8b","10":"sha256:631a3798b749fecc041a99929eb946618df723e15055e805ff752a1a1273481c","11":"sha256:1386b0fcb731d843f15fb64532f8b676c927821d69dd3d4503c973c3e2a04216","12":"sha256:9014f384de5f9a0b7418d5869ad349abb9588d16bd09ed650a163c045315dbff","13":"sha256:df606f3b71d4376d1a2108c09f0d3dab455fc30bcb67c60e91590c105e9025bf","14":"sha256:79f99fd6cce984287932edf0d009660bb488d663081f3d62ec3b23bc8bfbf6c2","15":"sha256:8139ed65c0a0a4b0f253b715c11cc52be027efe8a4774da9ccce35c78ef439da","16":"sha256:9724d2036305cbd729e1f484c5bad89971de977fff8a6723fef1873858dd1123","17":"registry.redhat.io/redhat/community-operator-index:v4.18","18":"sha256:3c1edf52f70bf9b1d1457e0c4111bc79cdaa1edd659ddbdb9d8176eff8b46956","19":"sha256:61664aa69b33349cc6de45e44ae6033e7f483c034ea01c0d9a8ca08a12d88e3a","2":"sha256:f1ca78c423f43f89a0411e40393642f64e4f8df9e5f61c25e31047c4cce170f9","20":"sha256:f042fa25014f3d37f3ea967d21f361d2a11833ae18f2c750318101b25d2497ce","21":"sha256:10e57ca7611f79710f05777dc6a8f31c7e04eb09da4d8d793a5acfbf0e4692d7","22":"sha256:eb928c13a46d3fb45f4a881892d023a92d610a5430be0ffd916aaf8da8e7d297","23":"sha256:eddedae7578d79b5a3f748000ae5c00b9f14a04710f9f9ec7b52fc569be5dfb8","24":"sha256:1f952cec1e5332b84bdffa249cd426f39087058d6544ddcec650a414c15a9b68","25":"registry.redhat.io/redhat/redhat-operator-index:v4.18","26":"sha256:86af77350cfe6fd69280157e4162aa0147873d9431c641ae4ad3e881ff768a73","27":"sha256:33a20002692769235e95271ab071783c57ff50681088fa1035b86af31e73cf20","28":"sha256:b472823604757237c2d16bd6f6221f4cf562aa3b05942c7f602e1e8b2e55a7c6","29":"registry.redhat.io/redhat/redhat-marketplace-index:v4.18","3":"sha256:188637a52cafee61ec461e92fb0c605e28be325b9ac1f2ac8a37d68e97654718","30":"sha256:5451aa441e5b8d8689c032405d410c8049a849ef2edf77e5b6a5ce2838c6569b","31":"sha256:cfde59e48cd5dee3721f34d249cb119cc3259fd857965d34f9c7ed83b0c363a1","32":"sha256:4ecc5bac651ff1942865baee5159582e9602c89b47eeab18400a32abcba8f690","33":"sha256:8eabac819f289e29d75c7ab172d8124554849a47f0b00770928c3eb19a5a31c4","34":"sha256:3b8d91a25eeb9f02041e947adb3487da3e7ab8449d3d2ad015827e7954df7b34","35":"sha256:a824e468cf8dd61d347e35b2ee5bc2f815666957647098e21a1bb56ff613e5b9","36":"registry.redhat.io/redhat/certified-operator-index:v4.18","37":"sha256:c3a77aa4d03b89ea284e3467a268e5989a77a2ef63e685eb1d5c5ea5b3922b7a","38":"sha256:9432c13d76bd4ba4eb9197c050cf88c0d701fa2055eeb59257e2e23901f9fdff","39":"sha256:8a3d37aa7a22c68afa963ecfb4b43c52cccf152580cd66e4d5382fb69e4037cc","4":"sha256:0f43c31aa3359159d4557dad3cfaf812d8ce44db9cb9ae970e06d3479070b660","40":"sha256:3f0aa9cd04713acc5c6fea721bd849e1500da8ae945e0b32000887f34d786e0b","41":"sha256:aa24edce3d740f84c40018e94cdbf2bc7375268d13d57c2d664e43a46ccea3fc","42":"sha256:8c885ea0b3c5124989f0a9b93eba98eb9fca6bbd0262772d85d90bf713a4d572","43":"sha256:b00c658332d6c6786bd969b26097c20a78c79c045f1692a8809234f5fb586c22","44":"sha256:7664a2d4cb10e82ed32abbf95799f43fc3d10135d7dd94799730de504a89680a","45":"sha256:916566bb9d0143352324233d460ad94697719c11c8c9158e3aea8f475941751f","46":"sha256:2632d7f05d5a992e91038ded81c715898f3fe803420a9b67a0201e9fd8075213","47":"sha256:c416b201d480bddb5a4960ec42f4740761a1335001cf84ba5ae19ad6857771b1","48":"sha256:97d26892192b552c16527bf2771e1b86528ab581a02dd9279cdf71c194830e3e","5":"sha256:01d2e67fd74086da701c39dac5b821822351cb0151f9afe72821c05df19953ad","6":"sha256:f4d4282cb53325e737ad68abbfcb70687ae04fb50353f4f0ba0ba5703b15009a","7":"sha256:6d5891cdd7dcf7c9081de8b364b4c96446b7f946f7880fbae291a4592a198264","8":"sha256:7e438b814f8e16f00b3fc4b69991af80eee79ae111d2a707f34aa64b2ccbb6eb","9":"sha256:91af633e585621630c40d14f188e37d36b44678d0a59e582d850bf8d593d3a0c"},"containers":{"2025-12":{"0":9,"1":8,"10":2,"11":2,"12":2,"13":2,"14":2,"15":2,"16":2,"17":1,"18":1,"19":1,"2":7,"20":1,"21":1,"22":1,"23":1,"24":1,"25":1,"26":1,"27":1,"28":1,"29":1,"3":6,"30":1,"31":1,"32":1,"33":1,"34":1,"35":1,"36":1,"37":1,"38":1,"39":1,"4":5,"40":1,"41":1,"42":1,"43":1,"44":1,"45":1,"46":1,"47":1,"48":1,"5":4,"6":3,"7":2,"8":2,"9":2}}}operator-controller-controller-manager-7cbd59c7f8-qcz9t.json0000640000000000000000000002474515114274061032471 0ustar0000000000000000config/pod/openshift-operator-controller{"metadata":{"name":"operator-controller-controller-manager-7cbd59c7f8-qcz9t","generateName":"operator-controller-controller-manager-7cbd59c7f8-","namespace":"openshift-operator-controller","uid":"1897bad8-3875-4bd6-b877-b1abe1fe76c7","resourceVersion":"11958","creationTimestamp":"2025-12-04T11:38:18Z","labels":{"control-plane":"operator-controller-controller-manager","pod-template-hash":"7cbd59c7f8"},"annotations":{"k8s.ovn.org/pod-networks":"{\"default\":{\"ip_addresses\":[\"10.128.0.41/23\"],\"mac_address\":\"0a:58:0a:80:00:29\",\"gateway_ips\":[\"10.128.0.1\"],\"routes\":[{\"dest\":\"10.128.0.0/16\",\"nextHop\":\"10.128.0.1\"},{\"dest\":\"172.30.0.0/16\",\"nextHop\":\"10.128.0.1\"},{\"dest\":\"169.254.0.5/32\",\"nextHop\":\"10.128.0.1\"},{\"dest\":\"100.64.0.0/16\",\"nextHop\":\"10.128.0.1\"}],\"ip_address\":\"10.128.0.41/23\",\"gateway_ip\":\"10.128.0.1\",\"role\":\"primary\"}}","k8s.v1.cni.cncf.io/network-status":"[{\n \"name\": \"ovn-kubernetes\",\n \"interface\": \"eth0\",\n \"ips\": [\n \"10.128.0.41\"\n ],\n \"mac\": \"0a:58:0a:80:00:29\",\n \"default\": true,\n \"dns\": {}\n}]","kubectl.kubernetes.io/default-container":"manager","openshift.io/required-scc":"privileged","openshift.io/scc":"privileged","seccomp.security.alpha.kubernetes.io/pod":"runtime/default"},"ownerReferences":[{"apiVersion":"apps/v1","kind":"ReplicaSet","name":"operator-controller-controller-manager-7cbd59c7f8","uid":"ce025b1b-0d86-4adb-812c-c250c7c1635d","controller":true,"blockOwnerDeletion":true}]},"spec":{"volumes":[{"name":"cache","emptyDir":{}},{"name":"ca-certs","projected":{"sources":[{"configMap":{"name":"operator-controller-trusted-ca-bundle","items":[{"key":"ca-bundle.crt","path":"ca-bundle.crt"}],"optional":false}},{"configMap":{"name":"openshift-service-ca.crt","items":[{"key":"service-ca.crt","path":"service-ca.crt"}],"optional":false}}],"defaultMode":420}},{"name":"etc-containers","hostPath":{"path":"/etc/containers","type":"Directory"}},{"name":"etc-docker","hostPath":{"path":"/etc/docker","type":"Directory"}},{"name":"kube-api-access-dkqp2","projected":{"sources":[{"serviceAccountToken":{"expirationSeconds":3607,"path":"token"}},{"configMap":{"name":"kube-root-ca.crt","items":[{"key":"ca.crt","path":"ca.crt"}]}},{"downwardAPI":{"items":[{"path":"namespace","fieldRef":{"apiVersion":"v1","fieldPath":"metadata.namespace"}}]}},{"configMap":{"name":"openshift-service-ca.crt","items":[{"key":"service-ca.crt","path":"service-ca.crt"}]}}],"defaultMode":420}}],"containers":[{"name":"manager","image":"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1f952cec1e5332b84bdffa249cd426f39087058d6544ddcec650a414c15a9b68","command":["/manager"],"args":["--health-probe-bind-address=:8081","--metrics-bind-address=127.0.0.1:8080","--leader-elect","--catalogd-cas-dir=/var/ca-certs","--v=2","--global-pull-secret=openshift-config/pull-secret"],"env":[{"name":"SSL_CERT_DIR","value":"/var/ca-certs"}],"resources":{"requests":{"cpu":"10m","memory":"64Mi"}},"volumeMounts":[{"name":"cache","mountPath":"/var/cache"},{"name":"ca-certs","readOnly":true,"mountPath":"/var/ca-certs"},{"name":"etc-containers","readOnly":true,"mountPath":"/etc/containers"},{"name":"etc-docker","readOnly":true,"mountPath":"/etc/docker"},{"name":"kube-api-access-dkqp2","readOnly":true,"mountPath":"/var/run/secrets/kubernetes.io/serviceaccount"}],"livenessProbe":{"httpGet":{"path":"/healthz","port":8081,"scheme":"HTTP"},"initialDelaySeconds":15,"timeoutSeconds":1,"periodSeconds":20,"successThreshold":1,"failureThreshold":3},"readinessProbe":{"httpGet":{"path":"/readyz","port":8081,"scheme":"HTTP"},"initialDelaySeconds":5,"timeoutSeconds":1,"periodSeconds":10,"successThreshold":1,"failureThreshold":3},"terminationMessagePath":"/dev/termination-log","terminationMessagePolicy":"FallbackToLogsOnError","imagePullPolicy":"IfNotPresent","securityContext":{"capabilities":{"drop":["ALL"]},"allowPrivilegeEscalation":false}},{"name":"kube-rbac-proxy","image":"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69ffd8f8dcceedc2d6eb306cea33f8beabc1be1308cd5f4ee8b9a8e3eab9843","args":["--secure-listen-address=0.0.0.0:8443","--http2-disable","--upstream=http://127.0.0.1:8080/","--logtostderr=true","--v=2"],"ports":[{"name":"https","containerPort":8443,"protocol":"TCP"}],"resources":{"requests":{"cpu":"5m","memory":"64Mi"}},"volumeMounts":[{"name":"kube-api-access-dkqp2","readOnly":true,"mountPath":"/var/run/secrets/kubernetes.io/serviceaccount"}],"terminationMessagePath":"/dev/termination-log","terminationMessagePolicy":"FallbackToLogsOnError","imagePullPolicy":"IfNotPresent","securityContext":{"capabilities":{"drop":["ALL"]},"allowPrivilegeEscalation":false}}],"restartPolicy":"Always","terminationGracePeriodSeconds":10,"dnsPolicy":"ClusterFirst","nodeSelector":{"kubernetes.io/os":"linux","node-role.kubernetes.io/master":""},"serviceAccountName":"operator-controller-controller-manager","serviceAccount":"operator-controller-controller-manager","nodeName":"master-0","securityContext":{"seLinuxOptions":{"type":"spc_t"},"runAsNonRoot":true,"seccompProfile":{"type":"RuntimeDefault"}},"affinity":{"nodeAffinity":{"requiredDuringSchedulingIgnoredDuringExecution":{"nodeSelectorTerms":[{"matchExpressions":[{"key":"kubernetes.io/arch","operator":"In","values":["amd64","arm64","ppc64le","s390x"]},{"key":"kubernetes.io/os","operator":"In","values":["linux"]}]}]}}},"schedulerName":"default-scheduler","tolerations":[{"key":"node-role.kubernetes.io/master","operator":"Exists","effect":"NoSchedule"},{"key":"node.kubernetes.io/unreachable","operator":"Exists","effect":"NoExecute","tolerationSeconds":120},{"key":"node.kubernetes.io/not-ready","operator":"Exists","effect":"NoExecute","tolerationSeconds":120},{"key":"node.kubernetes.io/memory-pressure","operator":"Exists","effect":"NoSchedule"}],"priorityClassName":"system-cluster-critical","priority":2000000000,"enableServiceLinks":true,"preemptionPolicy":"PreemptLowerPriority"},"status":{"phase":"Running","conditions":[{"type":"PodReadyToStartContainers","status":"True","lastProbeTime":null,"lastTransitionTime":"2025-12-04T11:38:22Z"},{"type":"Initialized","status":"True","lastProbeTime":null,"lastTransitionTime":"2025-12-04T11:38:18Z"},{"type":"Ready","status":"True","lastProbeTime":null,"lastTransitionTime":"2025-12-04T11:59:30Z"},{"type":"ContainersReady","status":"True","lastProbeTime":null,"lastTransitionTime":"2025-12-04T11:59:30Z"},{"type":"PodScheduled","status":"True","lastProbeTime":null,"lastTransitionTime":"2025-12-04T11:38:18Z"}],"hostIP":"192.168.32.10","hostIPs":[{"ip":"192.168.32.10"}],"podIP":"10.128.0.41","podIPs":[{"ip":"10.128.0.41"}],"startTime":"2025-12-04T11:38:18Z","containerStatuses":[{"name":"kube-rbac-proxy","state":{"running":{"startedAt":"2025-12-04T11:38:21Z"}},"lastState":{},"ready":true,"restartCount":0,"image":"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69ffd8f8dcceedc2d6eb306cea33f8beabc1be1308cd5f4ee8b9a8e3eab9843","imageID":"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69ffd8f8dcceedc2d6eb306cea33f8beabc1be1308cd5f4ee8b9a8e3eab9843","containerID":"cri-o://2ff1bb7b9640a70f24f91de11e060c8d7d8ad652495e8fa6de3fe55f6ad72da8","started":true,"volumeMounts":[{"name":"kube-api-access-dkqp2","mountPath":"/var/run/secrets/kubernetes.io/serviceaccount","readOnly":true,"recursiveReadOnly":"Disabled"}]},{"name":"manager","state":{"running":{"startedAt":"2025-12-04T11:52:43Z"}},"lastState":{"terminated":{"exitCode":1,"reason":"Error","message":"ork.io\" controllerKind=\"ClusterCatalog\" ClusterCatalog=\"openshift-redhat-marketplace\" namespace=\"\" name=\"openshift-redhat-marketplace\" reconcileID=\"297f1921-5b46-4bea-a4d2-b4566cc3df0b\"\nI1204 11:48:04.898441 1 clustercatalog_controller.go:90] \"reconcile ending\" logger=\"cluster-catalog\" controller=\"clustercatalog\" controllerGroup=\"olm.operatorframework.io\" controllerKind=\"ClusterCatalog\" ClusterCatalog=\"openshift-redhat-marketplace\" namespace=\"\" name=\"openshift-redhat-marketplace\" reconcileID=\"297f1921-5b46-4bea-a4d2-b4566cc3df0b\"\nI1204 11:48:04.898558 1 clustercatalog_controller.go:54] \"reconcile starting\" logger=\"cluster-catalog\" controller=\"clustercatalog\" controllerGroup=\"olm.operatorframework.io\" controllerKind=\"ClusterCatalog\" ClusterCatalog=\"openshift-redhat-operators\" namespace=\"\" name=\"openshift-redhat-operators\" reconcileID=\"ff3f7453-1ef5-4dd1-8fe4-742273763c6e\"\nI1204 11:48:08.158279 1 clustercatalog_controller.go:90] \"reconcile ending\" logger=\"cluster-catalog\" controller=\"clustercatalog\" controllerGroup=\"olm.operatorframework.io\" controllerKind=\"ClusterCatalog\" ClusterCatalog=\"openshift-redhat-operators\" namespace=\"\" name=\"openshift-redhat-operators\" reconcileID=\"ff3f7453-1ef5-4dd1-8fe4-742273763c6e\"\nE1204 11:51:29.691923 1 leaderelection.go:429] Failed to update lock optimitically: Timeout: request did not complete within requested timeout - context deadline exceeded, falling back to slow path\nE1204 11:52:29.695018 1 leaderelection.go:436] error retrieving resource lock openshift-operator-controller/9c4404e7.operatorframework.io: the server was unable to return a response in the time allotted, but may still be processing the request (get leases.coordination.k8s.io 9c4404e7.operatorframework.io)\nI1204 11:52:42.685769 1 leaderelection.go:297] failed to renew lease openshift-operator-controller/9c4404e7.operatorframework.io: timed out waiting for the condition\nE1204 11:52:42.685948 1 main.go:362] \"problem running manager\" err=\"leader election lost\" logger=\"setup\"\n","startedAt":"2025-12-04T11:41:38Z","finishedAt":"2025-12-04T11:52:42Z","containerID":"cri-o://87d19ac7ca9a3b1f2ae18249fc9bca976ae2df5c1da057ed782ff5452aa064ed"}},"ready":true,"restartCount":2,"image":"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1f952cec1e5332b84bdffa249cd426f39087058d6544ddcec650a414c15a9b68","imageID":"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1f952cec1e5332b84bdffa249cd426f39087058d6544ddcec650a414c15a9b68","containerID":"cri-o://516af80a6b1543d560759a7b981a4cb5acdf80045815ece9d76bdd6e0c80bd3b","started":true,"volumeMounts":[{"name":"cache","mountPath":"/var/cache"},{"name":"ca-certs","mountPath":"/var/ca-certs","readOnly":true,"recursiveReadOnly":"Disabled"},{"name":"etc-containers","mountPath":"/etc/containers","readOnly":true,"recursiveReadOnly":"Disabled"},{"name":"etc-docker","mountPath":"/etc/docker","readOnly":true,"recursiveReadOnly":"Disabled"},{"name":"kube-api-access-dkqp2","mountPath":"/var/run/secrets/kubernetes.io/serviceaccount","readOnly":true,"recursiveReadOnly":"Disabled"}]}],"qosClass":"Burstable"}}config/pod/openshift-network-node-identity/network-node-identity-f8hvq.json0000640000000000000000000002462515114274061025731 0ustar0000000000000000{"metadata":{"name":"network-node-identity-f8hvq","generateName":"network-node-identity-","namespace":"openshift-network-node-identity","uid":"7203ec93-6572-4032-825c-140029a86058","resourceVersion":"11248","creationTimestamp":"2025-12-04T11:36:49Z","labels":{"app":"network-node-identity","component":"network","controller-revision-hash":"656c887c88","kubernetes.io/os":"linux","openshift.io/component":"network","pod-template-generation":"1","type":"infra"},"annotations":{"openshift.io/required-scc":"hostnetwork-v2","openshift.io/scc":"hostnetwork-v2","seccomp.security.alpha.kubernetes.io/pod":"runtime/default"},"ownerReferences":[{"apiVersion":"apps/v1","kind":"DaemonSet","name":"network-node-identity","uid":"70fd8a0b-a0c8-48c9-b30a-f23c8cbb02a4","controller":true,"blockOwnerDeletion":true}]},"spec":{"volumes":[{"name":"webhook-cert","secret":{"secretName":"network-node-identity-cert","defaultMode":420}},{"name":"env-overrides","configMap":{"name":"env-overrides","defaultMode":420,"optional":true}},{"name":"ovnkube-identity-cm","configMap":{"name":"ovnkube-identity-cm","items":[{"key":"additional-cert-acceptance-cond.json","path":"additional-cert-acceptance-cond.json"},{"key":"additional-pod-admission-cond.json","path":"additional-pod-admission-cond.json"}],"defaultMode":420}},{"name":"kube-api-access-gkcj4","projected":{"sources":[{"serviceAccountToken":{"expirationSeconds":3607,"path":"token"}},{"configMap":{"name":"kube-root-ca.crt","items":[{"key":"ca.crt","path":"ca.crt"}]}},{"downwardAPI":{"items":[{"path":"namespace","fieldRef":{"apiVersion":"v1","fieldPath":"metadata.namespace"}}]}},{"configMap":{"name":"openshift-service-ca.crt","items":[{"key":"service-ca.crt","path":"service-ca.crt"}]}}],"defaultMode":420}}],"containers":[{"name":"webhook","image":"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e3e65409fc2b27ad0aaeb500a39e264663d2980821f099b830b551785ce4ce8b","command":["/bin/bash","-c","set -xe\nif [[ -f \"/env/_master\" ]]; then\n set -o allexport\n source \"/env/_master\"\n set +o allexport\nfi\n# OVN-K will try to remove hybrid overlay node annotations even when the hybrid overlay is not enabled.\n# https://github.com/ovn-org/ovn-kubernetes/blob/ac6820df0b338a246f10f412cd5ec903bd234694/go-controller/pkg/ovn/master.go#L791\nho_enable=\"--enable-hybrid-overlay\"\necho \"I$(date \"+%m%d %H:%M:%S.%N\") - network-node-identity - start webhook\"\n# extra-allowed-user: service account `ovn-kubernetes-control-plane`\n# sets pod annotations in multi-homing layer3 network controller (cluster-manager)\nexec /usr/bin/ovnkube-identity --k8s-apiserver=https://api-int.sno.openstack.lab:6443 \\\n --webhook-cert-dir=\"/etc/webhook-cert\" \\\n --webhook-host=127.0.0.1 \\\n --webhook-port=9743 \\\n ${ho_enable} \\\n --enable-interconnect \\\n --disable-approver \\\n --extra-allowed-user=\"system:serviceaccount:openshift-ovn-kubernetes:ovn-kubernetes-control-plane\" \\\n --wait-for-kubernetes-api=200s \\\n --pod-admission-conditions=\"/var/run/ovnkube-identity-config/additional-pod-admission-cond.json\" \\\n --loglevel=\"${LOGLEVEL}\"\n"],"env":[{"name":"LOGLEVEL","value":"2"},{"name":"KUBERNETES_NODE_NAME","valueFrom":{"fieldRef":{"apiVersion":"v1","fieldPath":"spec.nodeName"}}}],"resources":{"requests":{"cpu":"10m","memory":"50Mi"}},"volumeMounts":[{"name":"webhook-cert","mountPath":"/etc/webhook-cert/"},{"name":"env-overrides","mountPath":"/env"},{"name":"ovnkube-identity-cm","mountPath":"/var/run/ovnkube-identity-config"},{"name":"kube-api-access-gkcj4","readOnly":true,"mountPath":"/var/run/secrets/kubernetes.io/serviceaccount"}],"terminationMessagePath":"/dev/termination-log","terminationMessagePolicy":"FallbackToLogsOnError","imagePullPolicy":"IfNotPresent","securityContext":{"capabilities":{"drop":["ALL"]},"runAsUser":1000550000,"runAsNonRoot":true,"allowPrivilegeEscalation":false}},{"name":"approver","image":"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e3e65409fc2b27ad0aaeb500a39e264663d2980821f099b830b551785ce4ce8b","command":["/bin/bash","-c","set -xe\nif [[ -f \"/env/_master\" ]]; then\n set -o allexport\n source \"/env/_master\"\n set +o allexport\nfi\n\necho \"I$(date \"+%m%d %H:%M:%S.%N\") - network-node-identity - start approver\"\nexec /usr/bin/ovnkube-identity --k8s-apiserver=https://api-int.sno.openstack.lab:6443 \\\n --disable-webhook \\\n --csr-acceptance-conditions=\"/var/run/ovnkube-identity-config/additional-cert-acceptance-cond.json\" \\\n --loglevel=\"${LOGLEVEL}\"\n"],"env":[{"name":"LOGLEVEL","value":"4"}],"resources":{"requests":{"cpu":"10m","memory":"50Mi"}},"volumeMounts":[{"name":"env-overrides","mountPath":"/env"},{"name":"ovnkube-identity-cm","mountPath":"/var/run/ovnkube-identity-config"},{"name":"kube-api-access-gkcj4","readOnly":true,"mountPath":"/var/run/secrets/kubernetes.io/serviceaccount"}],"terminationMessagePath":"/dev/termination-log","terminationMessagePolicy":"FallbackToLogsOnError","imagePullPolicy":"IfNotPresent","securityContext":{"capabilities":{"drop":["ALL"]},"runAsUser":1000550000,"runAsNonRoot":true,"allowPrivilegeEscalation":false}}],"restartPolicy":"Always","terminationGracePeriodSeconds":200,"dnsPolicy":"Default","nodeSelector":{"beta.kubernetes.io/os":"linux","node-role.kubernetes.io/master":""},"serviceAccountName":"network-node-identity","serviceAccount":"network-node-identity","nodeName":"master-0","hostNetwork":true,"securityContext":{"seLinuxOptions":{"level":"s0:c23,c22"},"supplementalGroups":[1000550000],"fsGroup":1000550000,"seccompProfile":{"type":"RuntimeDefault"}},"affinity":{"nodeAffinity":{"requiredDuringSchedulingIgnoredDuringExecution":{"nodeSelectorTerms":[{"matchFields":[{"key":"metadata.name","operator":"In","values":["master-0"]}]}]}}},"schedulerName":"default-scheduler","tolerations":[{"operator":"Exists"}],"priorityClassName":"system-node-critical","priority":2000001000,"enableServiceLinks":true,"preemptionPolicy":"PreemptLowerPriority"},"status":{"phase":"Running","conditions":[{"type":"PodReadyToStartContainers","status":"True","lastProbeTime":null,"lastTransitionTime":"2025-12-04T11:37:09Z"},{"type":"Initialized","status":"True","lastProbeTime":null,"lastTransitionTime":"2025-12-04T11:36:49Z"},{"type":"Ready","status":"True","lastProbeTime":null,"lastTransitionTime":"2025-12-04T11:51:59Z"},{"type":"ContainersReady","status":"True","lastProbeTime":null,"lastTransitionTime":"2025-12-04T11:51:59Z"},{"type":"PodScheduled","status":"True","lastProbeTime":null,"lastTransitionTime":"2025-12-04T11:36:49Z"}],"hostIP":"192.168.32.10","hostIPs":[{"ip":"192.168.32.10"}],"podIP":"192.168.32.10","podIPs":[{"ip":"192.168.32.10"}],"startTime":"2025-12-04T11:36:49Z","containerStatuses":[{"name":"approver","state":{"running":{"startedAt":"2025-12-04T11:51:58Z"}},"lastState":{"terminated":{"exitCode":1,"reason":"Error","message":"o\" controllerKind=\"CertificateSigningRequest\" worker count=1\nI1204 11:45:32.664202 1 approver.go:230] Finished syncing CSR csr-qfbn7 for unknown node in 251.269µs\nI1204 11:45:32.664484 1 approver.go:230] Finished syncing CSR csr-cz8bh for unknown node in 21.701µs\nE1204 11:51:27.890436 1 leaderelection.go:429] Failed to update lock optimistically: Put \"https://api-int.sno.openstack.lab:6443/apis/coordination.k8s.io/v1/namespaces/openshift-network-node-identity/leases/ovnkube-identity?timeout=15s\": context deadline exceeded (Client.Timeout exceeded while awaiting headers), falling back to slow path\nE1204 11:51:42.890275 1 leaderelection.go:436] error retrieving resource lock openshift-network-node-identity/ovnkube-identity: Get \"https://api-int.sno.openstack.lab:6443/apis/coordination.k8s.io/v1/namespaces/openshift-network-node-identity/leases/ovnkube-identity?timeout=15s\": context deadline exceeded\nI1204 11:51:42.890342 1 leaderelection.go:297] failed to renew lease openshift-network-node-identity/ovnkube-identity: context deadline exceeded\nE1204 11:51:57.890719 1 leaderelection.go:322] Failed to release lock: Put \"https://api-int.sno.openstack.lab:6443/apis/coordination.k8s.io/v1/namespaces/openshift-network-node-identity/leases/ovnkube-identity?timeout=15s\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)\nI1204 11:51:57.891006 1 internal.go:538] \"Stopping and waiting for non leader election runnables\"\nI1204 11:51:57.891089 1 internal.go:542] \"Stopping and waiting for leader election runnables\"\nF1204 11:51:57.891048 1 ovnkubeidentity.go:309] error running approver: leader election lost\nI1204 11:51:57.891125 1 internal.go:550] \"Stopping and waiting for caches\"\nI1204 11:51:57.891144 1 internal.go:554] \"Stopping and waiting for webhooks\"\nI1204 11:51:57.891164 1 internal.go:557] \"Stopping and waiting for HTTP servers\"\nI1204 11:51:57.891190 1 internal.go:561] \"Wait completed, proceeding to shutdown the manager\"\n","startedAt":"2025-12-04T11:40:30Z","finishedAt":"2025-12-04T11:51:57Z","containerID":"cri-o://6ee863793f57b0f18fae063e87f584d73fae0eb1eb8a1275618f4008ffa23d64"}},"ready":true,"restartCount":2,"image":"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e3e65409fc2b27ad0aaeb500a39e264663d2980821f099b830b551785ce4ce8b","imageID":"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e3e65409fc2b27ad0aaeb500a39e264663d2980821f099b830b551785ce4ce8b","containerID":"cri-o://813c42ef38b80f036a503096d52dfe1af020f6bc0a58ddcb44f7160546682e9a","started":true,"volumeMounts":[{"name":"env-overrides","mountPath":"/env"},{"name":"ovnkube-identity-cm","mountPath":"/var/run/ovnkube-identity-config"},{"name":"kube-api-access-gkcj4","mountPath":"/var/run/secrets/kubernetes.io/serviceaccount","readOnly":true,"recursiveReadOnly":"Disabled"}]},{"name":"webhook","state":{"running":{"startedAt":"2025-12-04T11:37:08Z"}},"lastState":{},"ready":true,"restartCount":0,"image":"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e3e65409fc2b27ad0aaeb500a39e264663d2980821f099b830b551785ce4ce8b","imageID":"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e3e65409fc2b27ad0aaeb500a39e264663d2980821f099b830b551785ce4ce8b","containerID":"cri-o://7dac80bd5755cc3e7a057bbde1ce5a77a32ead50c320b8fbd76da84cd7e15c2f","started":true,"volumeMounts":[{"name":"webhook-cert","mountPath":"/etc/webhook-cert/"},{"name":"env-overrides","mountPath":"/env"},{"name":"ovnkube-identity-cm","mountPath":"/var/run/ovnkube-identity-config"},{"name":"kube-api-access-gkcj4","mountPath":"/var/run/secrets/kubernetes.io/serviceaccount","readOnly":true,"recursiveReadOnly":"Disabled"}]}],"qosClass":"Burstable"}}config/pod/openshift-machine-api/machine-api-operator-88d48b57d-9fjtd.json0000640000000000000000000002376315114274061024627 0ustar0000000000000000{"metadata":{"name":"machine-api-operator-88d48b57d-9fjtd","generateName":"machine-api-operator-88d48b57d-","namespace":"openshift-machine-api","uid":"c50317d3-f7cd-4133-845e-44add57ac378","resourceVersion":"11260","creationTimestamp":"2025-12-04T11:38:45Z","labels":{"k8s-app":"machine-api-operator","pod-template-hash":"88d48b57d"},"annotations":{"capability.openshift.io/name":"MachineAPI","k8s.ovn.org/pod-networks":"{\"default\":{\"ip_addresses\":[\"10.128.0.66/23\"],\"mac_address\":\"0a:58:0a:80:00:42\",\"gateway_ips\":[\"10.128.0.1\"],\"routes\":[{\"dest\":\"10.128.0.0/16\",\"nextHop\":\"10.128.0.1\"},{\"dest\":\"172.30.0.0/16\",\"nextHop\":\"10.128.0.1\"},{\"dest\":\"169.254.0.5/32\",\"nextHop\":\"10.128.0.1\"},{\"dest\":\"100.64.0.0/16\",\"nextHop\":\"10.128.0.1\"}],\"ip_address\":\"10.128.0.66/23\",\"gateway_ip\":\"10.128.0.1\",\"role\":\"primary\"}}","k8s.v1.cni.cncf.io/network-status":"[{\n \"name\": \"ovn-kubernetes\",\n \"interface\": \"eth0\",\n \"ips\": [\n \"10.128.0.66\"\n ],\n \"mac\": \"0a:58:0a:80:00:42\",\n \"default\": true,\n \"dns\": {}\n}]","kubectl.kubernetes.io/default-container":"machine-api-operator","openshift.io/required-scc":"restricted-v2","openshift.io/scc":"restricted-v2","seccomp.security.alpha.kubernetes.io/pod":"runtime/default"},"ownerReferences":[{"apiVersion":"apps/v1","kind":"ReplicaSet","name":"machine-api-operator-88d48b57d","uid":"fb6894a3-fe31-47bd-b600-edd3340da088","controller":true,"blockOwnerDeletion":true}]},"spec":{"volumes":[{"name":"config","configMap":{"name":"kube-rbac-proxy","defaultMode":420}},{"name":"images","configMap":{"name":"machine-api-operator-images","defaultMode":420}},{"name":"machine-api-operator-tls","secret":{"secretName":"machine-api-operator-tls","defaultMode":420}},{"name":"kube-api-access-s2jw8","projected":{"sources":[{"serviceAccountToken":{"expirationSeconds":3607,"path":"token"}},{"configMap":{"name":"kube-root-ca.crt","items":[{"key":"ca.crt","path":"ca.crt"}]}},{"downwardAPI":{"items":[{"path":"namespace","fieldRef":{"apiVersion":"v1","fieldPath":"metadata.namespace"}}]}},{"configMap":{"name":"openshift-service-ca.crt","items":[{"key":"service-ca.crt","path":"service-ca.crt"}]}}],"defaultMode":420}}],"containers":[{"name":"kube-rbac-proxy","image":"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69ffd8f8dcceedc2d6eb306cea33f8beabc1be1308cd5f4ee8b9a8e3eab9843","args":["--secure-listen-address=0.0.0.0:8443","--upstream=http://localhost:8080/","--tls-cert-file=/etc/tls/private/tls.crt","--tls-private-key-file=/etc/tls/private/tls.key","--config-file=/etc/kube-rbac-proxy/config-file.yaml","--tls-cipher-suites=TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305,TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305","--logtostderr=true","--v=3"],"ports":[{"name":"https","containerPort":8443,"protocol":"TCP"}],"resources":{"requests":{"cpu":"10m","memory":"20Mi"}},"volumeMounts":[{"name":"config","mountPath":"/etc/kube-rbac-proxy"},{"name":"machine-api-operator-tls","mountPath":"/etc/tls/private"},{"name":"kube-api-access-s2jw8","readOnly":true,"mountPath":"/var/run/secrets/kubernetes.io/serviceaccount"}],"terminationMessagePath":"/dev/termination-log","terminationMessagePolicy":"FallbackToLogsOnError","imagePullPolicy":"IfNotPresent","securityContext":{"capabilities":{"drop":["ALL"]},"runAsUser":1000500000,"runAsNonRoot":true,"allowPrivilegeEscalation":false}},{"name":"machine-api-operator","image":"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7c2431a990bcddde98829abda81950247021a2ebbabc964b1516ea046b5f1d4e","command":["/machine-api-operator"],"args":["start","--images-json=/etc/machine-api-operator-config/images/images.json","--alsologtostderr","--v=3"],"env":[{"name":"RELEASE_VERSION","value":"4.18.29"},{"name":"COMPONENT_NAMESPACE","valueFrom":{"fieldRef":{"apiVersion":"v1","fieldPath":"metadata.namespace"}}},{"name":"POD_NAME","valueFrom":{"fieldRef":{"apiVersion":"v1","fieldPath":"metadata.name"}}},{"name":"METRICS_PORT","value":"8080"}],"resources":{"requests":{"cpu":"10m","memory":"50Mi"}},"volumeMounts":[{"name":"images","mountPath":"/etc/machine-api-operator-config/images"},{"name":"kube-api-access-s2jw8","readOnly":true,"mountPath":"/var/run/secrets/kubernetes.io/serviceaccount"}],"terminationMessagePath":"/dev/termination-log","terminationMessagePolicy":"FallbackToLogsOnError","imagePullPolicy":"IfNotPresent","securityContext":{"capabilities":{"drop":["ALL"]},"runAsUser":1000500000,"runAsNonRoot":true,"allowPrivilegeEscalation":false}}],"restartPolicy":"Always","terminationGracePeriodSeconds":30,"dnsPolicy":"ClusterFirst","nodeSelector":{"node-role.kubernetes.io/master":""},"serviceAccountName":"machine-api-operator","serviceAccount":"machine-api-operator","nodeName":"master-0","securityContext":{"seLinuxOptions":{"level":"s0:c22,c19"},"fsGroup":1000500000,"seccompProfile":{"type":"RuntimeDefault"}},"imagePullSecrets":[{"name":"machine-api-operator-dockercfg-tpbt2"}],"schedulerName":"default-scheduler","tolerations":[{"key":"node-role.kubernetes.io/master","operator":"Exists","effect":"NoSchedule"},{"key":"node.kubernetes.io/unreachable","operator":"Exists","effect":"NoExecute","tolerationSeconds":120},{"key":"node.kubernetes.io/not-ready","operator":"Exists","effect":"NoExecute","tolerationSeconds":120},{"key":"node.kubernetes.io/memory-pressure","operator":"Exists","effect":"NoSchedule"}],"priorityClassName":"system-node-critical","priority":2000001000,"enableServiceLinks":true,"preemptionPolicy":"PreemptLowerPriority"},"status":{"phase":"Running","conditions":[{"type":"PodReadyToStartContainers","status":"True","lastProbeTime":null,"lastTransitionTime":"2025-12-04T11:45:18Z"},{"type":"Initialized","status":"True","lastProbeTime":null,"lastTransitionTime":"2025-12-04T11:38:45Z"},{"type":"Ready","status":"True","lastProbeTime":null,"lastTransitionTime":"2025-12-04T11:55:48Z"},{"type":"ContainersReady","status":"True","lastProbeTime":null,"lastTransitionTime":"2025-12-04T11:55:48Z"},{"type":"PodScheduled","status":"True","lastProbeTime":null,"lastTransitionTime":"2025-12-04T11:38:45Z"}],"hostIP":"192.168.32.10","hostIPs":[{"ip":"192.168.32.10"}],"podIP":"10.128.0.66","podIPs":[{"ip":"10.128.0.66"}],"startTime":"2025-12-04T11:38:45Z","containerStatuses":[{"name":"kube-rbac-proxy","state":{"running":{"startedAt":"2025-12-04T11:45:11Z"}},"lastState":{},"ready":true,"restartCount":0,"image":"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69ffd8f8dcceedc2d6eb306cea33f8beabc1be1308cd5f4ee8b9a8e3eab9843","imageID":"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69ffd8f8dcceedc2d6eb306cea33f8beabc1be1308cd5f4ee8b9a8e3eab9843","containerID":"cri-o://98d20604fd8ab25d783000ee1ca78a071d7a0dba8eedff22a4aa8bd932bbf14d","started":true,"volumeMounts":[{"name":"config","mountPath":"/etc/kube-rbac-proxy"},{"name":"machine-api-operator-tls","mountPath":"/etc/tls/private"},{"name":"kube-api-access-s2jw8","mountPath":"/var/run/secrets/kubernetes.io/serviceaccount","readOnly":true,"recursiveReadOnly":"Disabled"}]},{"name":"machine-api-operator","state":{"running":{"startedAt":"2025-12-04T11:55:47Z"}},"lastState":{"terminated":{"exitCode":255,"reason":"Error","message":" Syncing status: re-syncing\nI1204 11:45:54.667095 1 request.go:632] Waited for 194.398606ms due to client-side throttling, not priority and fairness, request: PUT:https://172.30.0.1:443/apis/config.openshift.io/v1/clusteroperators/machine-api/status\nI1204 11:45:54.677613 1 sync.go:77] Provider is NoOp, skipping synchronisation\nI1204 11:45:54.867119 1 request.go:632] Waited for 189.387143ms due to client-side throttling, not priority and fairness, request: GET:https://172.30.0.1:443/apis/config.openshift.io/v1/clusteroperators/machine-api\nI1204 11:45:54.957357 1 status.go:99] Syncing status: available\nI1204 11:45:55.066882 1 request.go:632] Waited for 109.367011ms due to client-side throttling, not priority and fairness, request: PUT:https://172.30.0.1:443/apis/config.openshift.io/v1/clusteroperators/machine-api/status\nE1204 11:51:52.365703 1 leaderelection.go:429] Failed to update lock optimitically: Timeout: request did not complete within requested timeout - context deadline exceeded, falling back to slow path\nE1204 11:52:52.367161 1 leaderelection.go:436] error retrieving resource lock openshift-machine-api/machine-api-operator: the server was unable to return a response in the time allotted, but may still be processing the request (get leases.coordination.k8s.io machine-api-operator)\nE1204 11:54:26.375213 1 leaderelection.go:429] Failed to update lock optimitically: Timeout: request did not complete within requested timeout - context deadline exceeded, falling back to slow path\nE1204 11:55:18.362224 1 leaderelection.go:436] error retrieving resource lock openshift-machine-api/machine-api-operator: Get \"https://172.30.0.1:443/apis/coordination.k8s.io/v1/namespaces/openshift-machine-api/leases/machine-api-operator\": context deadline exceeded\nI1204 11:55:18.362281 1 leaderelection.go:297] failed to renew lease openshift-machine-api/machine-api-operator: timed out waiting for the condition\nF1204 11:55:46.779196 1 start.go:104] Leader election lost\n","startedAt":"2025-12-04T11:45:18Z","finishedAt":"2025-12-04T11:55:46Z","containerID":"cri-o://bf5e3c44e8b979c1b2a2791e8de129301f2196277f040ae6c9ab59c2dfb23a10"}},"ready":true,"restartCount":1,"image":"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7c2431a990bcddde98829abda81950247021a2ebbabc964b1516ea046b5f1d4e","imageID":"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7c2431a990bcddde98829abda81950247021a2ebbabc964b1516ea046b5f1d4e","containerID":"cri-o://c5abefdcb3f076c4d8fa607cbb52db521c6fb9369a69d197d46f4eb8ebdfe177","started":true,"volumeMounts":[{"name":"images","mountPath":"/etc/machine-api-operator-config/images"},{"name":"kube-api-access-s2jw8","mountPath":"/var/run/secrets/kubernetes.io/serviceaccount","readOnly":true,"recursiveReadOnly":"Disabled"}]}],"qosClass":"Burstable"}}config/pod/openshift-machine-api/control-plane-machine-set-operator-7df95c79b5-7w5lm.json0000640000000000000000000001762415114274061027525 0ustar0000000000000000{"metadata":{"name":"control-plane-machine-set-operator-7df95c79b5-7w5lm","generateName":"control-plane-machine-set-operator-7df95c79b5-","namespace":"openshift-machine-api","uid":"14bd3485-aa02-4984-b4a3-59f360b1b327","resourceVersion":"11253","creationTimestamp":"2025-12-04T11:38:32Z","labels":{"k8s-app":"control-plane-machine-set-operator","pod-template-hash":"7df95c79b5"},"annotations":{"k8s.ovn.org/pod-networks":"{\"default\":{\"ip_addresses\":[\"10.128.0.48/23\"],\"mac_address\":\"0a:58:0a:80:00:30\",\"gateway_ips\":[\"10.128.0.1\"],\"routes\":[{\"dest\":\"10.128.0.0/16\",\"nextHop\":\"10.128.0.1\"},{\"dest\":\"172.30.0.0/16\",\"nextHop\":\"10.128.0.1\"},{\"dest\":\"169.254.0.5/32\",\"nextHop\":\"10.128.0.1\"},{\"dest\":\"100.64.0.0/16\",\"nextHop\":\"10.128.0.1\"}],\"ip_address\":\"10.128.0.48/23\",\"gateway_ip\":\"10.128.0.1\",\"role\":\"primary\"}}","k8s.v1.cni.cncf.io/network-status":"[{\n \"name\": \"ovn-kubernetes\",\n \"interface\": \"eth0\",\n \"ips\": [\n \"10.128.0.48\"\n ],\n \"mac\": \"0a:58:0a:80:00:30\",\n \"default\": true,\n \"dns\": {}\n}]","openshift.io/required-scc":"restricted-v2","openshift.io/scc":"restricted-v2","seccomp.security.alpha.kubernetes.io/pod":"runtime/default"},"ownerReferences":[{"apiVersion":"apps/v1","kind":"ReplicaSet","name":"control-plane-machine-set-operator-7df95c79b5","uid":"c80aa5ea-3cfa-4daf-963c-6e4732f97894","controller":true,"blockOwnerDeletion":true}]},"spec":{"volumes":[{"name":"control-plane-machine-set-operator-tls","secret":{"secretName":"control-plane-machine-set-operator-tls","defaultMode":420}},{"name":"kube-api-access-nfspg","projected":{"sources":[{"serviceAccountToken":{"expirationSeconds":3607,"path":"token"}},{"configMap":{"name":"kube-root-ca.crt","items":[{"key":"ca.crt","path":"ca.crt"}]}},{"downwardAPI":{"items":[{"path":"namespace","fieldRef":{"apiVersion":"v1","fieldPath":"metadata.namespace"}}]}},{"configMap":{"name":"openshift-service-ca.crt","items":[{"key":"service-ca.crt","path":"service-ca.crt"}]}}],"defaultMode":420}}],"containers":[{"name":"control-plane-machine-set-operator","image":"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fd3e9f8f00a59bda7483ec7dc8a0ed602f9ca30e3d72b22072dbdf2819da3f61","command":["/manager"],"args":["-v=2","--leader-elect=true","--leader-elect-lease-duration=137s","--leader-elect-renew-deadline=107s","--leader-elect-retry-period=26s","--leader-elect-resource-namespace=openshift-machine-api"],"ports":[{"name":"https","containerPort":9443,"protocol":"TCP"}],"env":[{"name":"RELEASE_VERSION","value":"4.18.29"},{"name":"COMPONENT_NAMESPACE","valueFrom":{"fieldRef":{"apiVersion":"v1","fieldPath":"metadata.namespace"}}}],"resources":{"requests":{"cpu":"10m","memory":"50Mi"}},"volumeMounts":[{"name":"control-plane-machine-set-operator-tls","mountPath":"/tmp/k8s-webhook-server/serving-certs"},{"name":"kube-api-access-nfspg","readOnly":true,"mountPath":"/var/run/secrets/kubernetes.io/serviceaccount"}],"terminationMessagePath":"/dev/termination-log","terminationMessagePolicy":"FallbackToLogsOnError","imagePullPolicy":"IfNotPresent","securityContext":{"capabilities":{"drop":["ALL"]},"runAsUser":1000500000,"runAsNonRoot":true,"allowPrivilegeEscalation":false}}],"restartPolicy":"Always","terminationGracePeriodSeconds":30,"dnsPolicy":"ClusterFirst","nodeSelector":{"node-role.kubernetes.io/master":""},"serviceAccountName":"control-plane-machine-set-operator","serviceAccount":"control-plane-machine-set-operator","nodeName":"master-0","securityContext":{"seLinuxOptions":{"level":"s0:c22,c19"},"fsGroup":1000500000,"seccompProfile":{"type":"RuntimeDefault"}},"schedulerName":"default-scheduler","tolerations":[{"key":"node-role.kubernetes.io/master","operator":"Exists","effect":"NoSchedule"},{"key":"node.kubernetes.io/not-ready","operator":"Exists","effect":"NoExecute","tolerationSeconds":300},{"key":"node.kubernetes.io/unreachable","operator":"Exists","effect":"NoExecute","tolerationSeconds":300},{"key":"node.kubernetes.io/memory-pressure","operator":"Exists","effect":"NoSchedule"}],"priorityClassName":"system-cluster-critical","priority":2000000000,"enableServiceLinks":true,"preemptionPolicy":"PreemptLowerPriority"},"status":{"phase":"Running","conditions":[{"type":"PodReadyToStartContainers","status":"True","lastProbeTime":null,"lastTransitionTime":"2025-12-04T11:38:41Z"},{"type":"Initialized","status":"True","lastProbeTime":null,"lastTransitionTime":"2025-12-04T11:38:33Z"},{"type":"Ready","status":"True","lastProbeTime":null,"lastTransitionTime":"2025-12-04T11:53:24Z"},{"type":"ContainersReady","status":"True","lastProbeTime":null,"lastTransitionTime":"2025-12-04T11:53:24Z"},{"type":"PodScheduled","status":"True","lastProbeTime":null,"lastTransitionTime":"2025-12-04T11:38:33Z"}],"hostIP":"192.168.32.10","hostIPs":[{"ip":"192.168.32.10"}],"podIP":"10.128.0.48","podIPs":[{"ip":"10.128.0.48"}],"startTime":"2025-12-04T11:38:33Z","containerStatuses":[{"name":"control-plane-machine-set-operator","state":{"running":{"startedAt":"2025-12-04T11:53:23Z"}},"lastState":{"terminated":{"exitCode":1,"reason":"Error","message":"a1.Machine\"\nI1204 11:47:07.047581 1 controller.go:183] \"Starting Controller\" controller=\"controlplanemachinesetgenerator\"\nI1204 11:47:07.158212 1 controller.go:217] \"Starting workers\" controller=\"controlplanemachineset\" worker count=1\nI1204 11:47:07.158328 1 controller.go:170] \"Reconciling control plane machine set\" controller=\"controlplanemachineset\" reconcileID=\"a67f96a9-6901-4cb3-8b4b-a1ba56d694f7\" namespace=\"openshift-machine-api\" name=\"cluster\"\nI1204 11:47:07.158437 1 controller.go:178] \"No control plane machine set found, setting operator status available\" controller=\"controlplanemachineset\" reconcileID=\"a67f96a9-6901-4cb3-8b4b-a1ba56d694f7\" namespace=\"openshift-machine-api\" name=\"cluster\"\nI1204 11:47:07.158507 1 controller.go:184] \"Finished reconciling control plane machine set\" controller=\"controlplanemachineset\" reconcileID=\"a67f96a9-6901-4cb3-8b4b-a1ba56d694f7\" namespace=\"openshift-machine-api\" name=\"cluster\"\nI1204 11:47:07.160500 1 controller.go:217] \"Starting workers\" controller=\"controlplanemachinesetgenerator\" worker count=1\nE1204 11:51:35.165428 1 leaderelection.go:429] Failed to update lock optimitically: Timeout: request did not complete within requested timeout - context deadline exceeded, falling back to slow path\nE1204 11:52:35.169552 1 leaderelection.go:436] error retrieving resource lock openshift-machine-api/control-plane-machine-set-leader: the server was unable to return a response in the time allotted, but may still be processing the request (get leases.coordination.k8s.io control-plane-machine-set-leader)\nI1204 11:52:48.160936 1 leaderelection.go:297] failed to renew lease openshift-machine-api/control-plane-machine-set-leader: timed out waiting for the condition\nE1204 11:53:22.166078 1 leaderelection.go:322] Failed to release lock: Timeout: request did not complete within requested timeout - context deadline exceeded\nE1204 11:53:22.166206 1 main.go:233] \"problem running manager\" err=\"leader election lost\" logger=\"setup\"\n","startedAt":"2025-12-04T11:42:12Z","finishedAt":"2025-12-04T11:53:22Z","containerID":"cri-o://41aa69685919e7f8f8af95aebbe0bc4d44d4ceec0966f5cc37e23c82ff7f99c8"}},"ready":true,"restartCount":2,"image":"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fd3e9f8f00a59bda7483ec7dc8a0ed602f9ca30e3d72b22072dbdf2819da3f61","imageID":"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fd3e9f8f00a59bda7483ec7dc8a0ed602f9ca30e3d72b22072dbdf2819da3f61","containerID":"cri-o://50a247f3243f1bb031bb64b878cb6bfaf702184aa3baa530e6774c26f20acc72","started":true,"volumeMounts":[{"name":"control-plane-machine-set-operator-tls","mountPath":"/tmp/k8s-webhook-server/serving-certs"},{"name":"kube-api-access-nfspg","mountPath":"/var/run/secrets/kubernetes.io/serviceaccount","readOnly":true,"recursiveReadOnly":"Disabled"}]}],"qosClass":"Burstable"}}config/pod/openshift-machine-api/cluster-baremetal-operator-78f758c7b9-zgkh5.json0000640000000000000000000002451615114274061026170 0ustar0000000000000000{"metadata":{"name":"cluster-baremetal-operator-78f758c7b9-zgkh5","generateName":"cluster-baremetal-operator-78f758c7b9-","namespace":"openshift-machine-api","uid":"7e3160a9-11d1-4845-ba30-1a49ae7339a9","resourceVersion":"11258","creationTimestamp":"2025-12-04T11:38:42Z","labels":{"k8s-app":"cluster-baremetal-operator","pod-template-hash":"78f758c7b9"},"annotations":{"capability.openshift.io/name":"baremetal","include.release.openshift.io/self-managed-high-availability":"true","include.release.openshift.io/single-node-developer":"true","k8s.ovn.org/pod-networks":"{\"default\":{\"ip_addresses\":[\"10.128.0.57/23\"],\"mac_address\":\"0a:58:0a:80:00:39\",\"gateway_ips\":[\"10.128.0.1\"],\"routes\":[{\"dest\":\"10.128.0.0/16\",\"nextHop\":\"10.128.0.1\"},{\"dest\":\"172.30.0.0/16\",\"nextHop\":\"10.128.0.1\"},{\"dest\":\"169.254.0.5/32\",\"nextHop\":\"10.128.0.1\"},{\"dest\":\"100.64.0.0/16\",\"nextHop\":\"10.128.0.1\"}],\"ip_address\":\"10.128.0.57/23\",\"gateway_ip\":\"10.128.0.1\",\"role\":\"primary\"}}","k8s.v1.cni.cncf.io/network-status":"[{\n \"name\": \"ovn-kubernetes\",\n \"interface\": \"eth0\",\n \"ips\": [\n \"10.128.0.57\"\n ],\n \"mac\": \"0a:58:0a:80:00:39\",\n \"default\": true,\n \"dns\": {}\n}]","openshift.io/required-scc":"anyuid","openshift.io/scc":"anyuid"},"ownerReferences":[{"apiVersion":"apps/v1","kind":"ReplicaSet","name":"cluster-baremetal-operator-78f758c7b9","uid":"a85499a8-b7d4-4de0-98e2-12b8c6ec328b","controller":true,"blockOwnerDeletion":true}]},"spec":{"volumes":[{"name":"cert","secret":{"secretName":"cluster-baremetal-webhook-server-cert","defaultMode":420}},{"name":"config","configMap":{"name":"baremetal-kube-rbac-proxy","defaultMode":420}},{"name":"cluster-baremetal-operator-tls","secret":{"secretName":"cluster-baremetal-operator-tls","defaultMode":420}},{"name":"images","configMap":{"name":"cluster-baremetal-operator-images","defaultMode":420}},{"name":"kube-api-access-fksx7","projected":{"sources":[{"serviceAccountToken":{"expirationSeconds":3607,"path":"token"}},{"configMap":{"name":"kube-root-ca.crt","items":[{"key":"ca.crt","path":"ca.crt"}]}},{"downwardAPI":{"items":[{"path":"namespace","fieldRef":{"apiVersion":"v1","fieldPath":"metadata.namespace"}}]}},{"configMap":{"name":"openshift-service-ca.crt","items":[{"key":"service-ca.crt","path":"service-ca.crt"}]}}],"defaultMode":420}}],"containers":[{"name":"cluster-baremetal-operator","image":"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a92c310ce30dcb3de85d6aac868e0d80919670fa29ef83d55edd96b0cae35563","command":["/usr/bin/cluster-baremetal-operator"],"args":["--enable-leader-election"],"ports":[{"name":"webhook-server","containerPort":9443,"protocol":"TCP"}],"env":[{"name":"RELEASE_VERSION","value":"4.18.29"},{"name":"COMPONENT_NAMESPACE","valueFrom":{"fieldRef":{"apiVersion":"v1","fieldPath":"metadata.namespace"}}},{"name":"METRICS_PORT","value":"8080"}],"resources":{"requests":{"cpu":"10m","memory":"50Mi"}},"volumeMounts":[{"name":"cert","readOnly":true,"mountPath":"/etc/cluster-baremetal-operator/tls"},{"name":"images","readOnly":true,"mountPath":"/etc/cluster-baremetal-operator/images"},{"name":"kube-api-access-fksx7","readOnly":true,"mountPath":"/var/run/secrets/kubernetes.io/serviceaccount"}],"terminationMessagePath":"/dev/termination-log","terminationMessagePolicy":"FallbackToLogsOnError","imagePullPolicy":"IfNotPresent","securityContext":{"capabilities":{"drop":["MKNOD"]}}},{"name":"baremetal-kube-rbac-proxy","image":"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69ffd8f8dcceedc2d6eb306cea33f8beabc1be1308cd5f4ee8b9a8e3eab9843","args":["--secure-listen-address=0.0.0.0:8443","--upstream=http://localhost:8080/","--tls-cert-file=/etc/tls/private/tls.crt","--tls-private-key-file=/etc/tls/private/tls.key","--tls-cipher-suites=TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305,TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305","--config-file=/etc/baremetal-kube-rbac-proxy/config-file.yaml","--logtostderr=true","--v=10"],"ports":[{"name":"https","containerPort":8443,"protocol":"TCP"}],"resources":{"requests":{"cpu":"10m","memory":"20Mi"}},"volumeMounts":[{"name":"config","mountPath":"/etc/baremetal-kube-rbac-proxy"},{"name":"cluster-baremetal-operator-tls","mountPath":"/etc/tls/private"},{"name":"kube-api-access-fksx7","readOnly":true,"mountPath":"/var/run/secrets/kubernetes.io/serviceaccount"}],"terminationMessagePath":"/dev/termination-log","terminationMessagePolicy":"FallbackToLogsOnError","imagePullPolicy":"IfNotPresent","securityContext":{"capabilities":{"drop":["MKNOD"]}}}],"restartPolicy":"Always","terminationGracePeriodSeconds":30,"dnsPolicy":"ClusterFirst","nodeSelector":{"node-role.kubernetes.io/master":""},"serviceAccountName":"cluster-baremetal-operator","serviceAccount":"cluster-baremetal-operator","nodeName":"master-0","securityContext":{"seLinuxOptions":{"level":"s0:c22,c19"},"runAsUser":65534,"runAsNonRoot":true},"imagePullSecrets":[{"name":"cluster-baremetal-operator-dockercfg-hkjv7"}],"schedulerName":"default-scheduler","tolerations":[{"key":"node-role.kubernetes.io/master","operator":"Exists","effect":"NoSchedule"},{"key":"node.kubernetes.io/unreachable","operator":"Exists","effect":"NoExecute","tolerationSeconds":120},{"key":"node.kubernetes.io/not-ready","operator":"Exists","effect":"NoExecute","tolerationSeconds":120},{"key":"node.kubernetes.io/memory-pressure","operator":"Exists","effect":"NoSchedule"}],"priorityClassName":"system-node-critical","priority":2000001000,"enableServiceLinks":true,"preemptionPolicy":"PreemptLowerPriority"},"status":{"phase":"Running","conditions":[{"type":"PodReadyToStartContainers","status":"True","lastProbeTime":null,"lastTransitionTime":"2025-12-04T11:39:23Z"},{"type":"Initialized","status":"True","lastProbeTime":null,"lastTransitionTime":"2025-12-04T11:38:42Z"},{"type":"Ready","status":"True","lastProbeTime":null,"lastTransitionTime":"2025-12-04T11:54:22Z"},{"type":"ContainersReady","status":"True","lastProbeTime":null,"lastTransitionTime":"2025-12-04T11:54:22Z"},{"type":"PodScheduled","status":"True","lastProbeTime":null,"lastTransitionTime":"2025-12-04T11:38:42Z"}],"hostIP":"192.168.32.10","hostIPs":[{"ip":"192.168.32.10"}],"podIP":"10.128.0.57","podIPs":[{"ip":"10.128.0.57"}],"startTime":"2025-12-04T11:38:42Z","containerStatuses":[{"name":"baremetal-kube-rbac-proxy","state":{"running":{"startedAt":"2025-12-04T11:39:22Z"}},"lastState":{},"ready":true,"restartCount":0,"image":"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69ffd8f8dcceedc2d6eb306cea33f8beabc1be1308cd5f4ee8b9a8e3eab9843","imageID":"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69ffd8f8dcceedc2d6eb306cea33f8beabc1be1308cd5f4ee8b9a8e3eab9843","containerID":"cri-o://aa2ca6d7bed669be922ad1ba5a0a322f3c44a0ef60988942199ce95ffc405be9","started":true,"volumeMounts":[{"name":"config","mountPath":"/etc/baremetal-kube-rbac-proxy"},{"name":"cluster-baremetal-operator-tls","mountPath":"/etc/tls/private"},{"name":"kube-api-access-fksx7","mountPath":"/var/run/secrets/kubernetes.io/serviceaccount","readOnly":true,"recursiveReadOnly":"Disabled"}]},{"name":"cluster-baremetal-operator","state":{"running":{"startedAt":"2025-12-04T11:54:21Z"}},"lastState":{"terminated":{"exitCode":1,"reason":"Error","message":"\"=\"Starting EventSource\" \"controller\"=\"provisioning\" \"controllerGroup\"=\"metal3.io\" \"controllerKind\"=\"Provisioning\" \"source\"=\"kind source: *v1.DaemonSet\"\nI1204 11:45:56.008030 1 controller.go:173] \"msg\"=\"Starting EventSource\" \"controller\"=\"provisioning\" \"controllerGroup\"=\"metal3.io\" \"controllerKind\"=\"Provisioning\" \"source\"=\"kind source: *v1.ClusterOperator\"\nI1204 11:45:56.008041 1 controller.go:173] \"msg\"=\"Starting EventSource\" \"controller\"=\"provisioning\" \"controllerGroup\"=\"metal3.io\" \"controllerKind\"=\"Provisioning\" \"source\"=\"kind source: *v1.Proxy\"\nI1204 11:45:56.008052 1 controller.go:173] \"msg\"=\"Starting EventSource\" \"controller\"=\"provisioning\" \"controllerGroup\"=\"metal3.io\" \"controllerKind\"=\"Provisioning\" \"source\"=\"kind source: *v1beta1.Machine\"\nI1204 11:45:56.008061 1 controller.go:181] \"msg\"=\"Starting Controller\" \"controller\"=\"provisioning\" \"controllerGroup\"=\"metal3.io\" \"controllerKind\"=\"Provisioning\"\nI1204 11:45:56.225289 1 controller.go:215] \"msg\"=\"Starting workers\" \"controller\"=\"provisioning\" \"controllerGroup\"=\"metal3.io\" \"controllerKind\"=\"Provisioning\" \"worker count\"=1\nE1204 11:51:42.850191 1 leaderelection.go:340] Failed to update lock optimitically: Timeout: request did not complete within requested timeout - context deadline exceeded, falling back to slow path\nE1204 11:52:42.851881 1 leaderelection.go:347] error retrieving resource lock openshift-machine-api/cluster-baremetal-operator: the server was unable to return a response in the time allotted, but may still be processing the request (get leases.coordination.k8s.io cluster-baremetal-operator)\nI1204 11:52:55.843671 1 leaderelection.go:285] failed to renew lease openshift-machine-api/cluster-baremetal-operator: timed out waiting for the condition\nE1204 11:53:29.848314 1 leaderelection.go:308] Failed to release lock: Timeout: request did not complete within requested timeout - context deadline exceeded\nE1204 11:53:29.848417 1 main.go:182] \"problem running manager\" err=\"leader election lost\"\n","startedAt":"2025-12-04T11:45:02Z","finishedAt":"2025-12-04T11:53:29Z","containerID":"cri-o://fb9fc7d90e1a7f1b9941af4021fef36a6cf8ea28ed08b7fb03d0f413775fbd30"}},"ready":true,"restartCount":4,"image":"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a92c310ce30dcb3de85d6aac868e0d80919670fa29ef83d55edd96b0cae35563","imageID":"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a92c310ce30dcb3de85d6aac868e0d80919670fa29ef83d55edd96b0cae35563","containerID":"cri-o://32e66f745e18dc99b4c4e9785252ea6bac8b15ebf03d317797384ca1ae5092e7","started":true,"volumeMounts":[{"name":"cert","mountPath":"/etc/cluster-baremetal-operator/tls","readOnly":true,"recursiveReadOnly":"Disabled"},{"name":"images","mountPath":"/etc/cluster-baremetal-operator/images","readOnly":true,"recursiveReadOnly":"Disabled"},{"name":"kube-api-access-fksx7","mountPath":"/var/run/secrets/kubernetes.io/serviceaccount","readOnly":true,"recursiveReadOnly":"Disabled"}]}],"qosClass":"Burstable"}}config/pod/openshift-machine-api/cluster-autoscaler-operator-5f49d774cd-894dk.json0000640000000000000000000002511215114274061026251 0ustar0000000000000000{"metadata":{"name":"cluster-autoscaler-operator-5f49d774cd-894dk","generateName":"cluster-autoscaler-operator-5f49d774cd-","namespace":"openshift-machine-api","uid":"e7fc7c16-5bca-49e5-aff0-7a8f80c6b639","resourceVersion":"11261","creationTimestamp":"2025-12-04T11:38:43Z","labels":{"k8s-app":"cluster-autoscaler-operator","pod-template-hash":"5f49d774cd"},"annotations":{"k8s.ovn.org/pod-networks":"{\"default\":{\"ip_addresses\":[\"10.128.0.59/23\"],\"mac_address\":\"0a:58:0a:80:00:3b\",\"gateway_ips\":[\"10.128.0.1\"],\"routes\":[{\"dest\":\"10.128.0.0/16\",\"nextHop\":\"10.128.0.1\"},{\"dest\":\"172.30.0.0/16\",\"nextHop\":\"10.128.0.1\"},{\"dest\":\"169.254.0.5/32\",\"nextHop\":\"10.128.0.1\"},{\"dest\":\"100.64.0.0/16\",\"nextHop\":\"10.128.0.1\"}],\"ip_address\":\"10.128.0.59/23\",\"gateway_ip\":\"10.128.0.1\",\"role\":\"primary\"}}","k8s.v1.cni.cncf.io/network-status":"[{\n \"name\": \"ovn-kubernetes\",\n \"interface\": \"eth0\",\n \"ips\": [\n \"10.128.0.59\"\n ],\n \"mac\": \"0a:58:0a:80:00:3b\",\n \"default\": true,\n \"dns\": {}\n}]","kubectl.kubernetes.io/default-container":"cluster-autoscaler-operator","openshift.io/required-scc":"restricted-v2","openshift.io/scc":"restricted-v2","seccomp.security.alpha.kubernetes.io/pod":"runtime/default"},"ownerReferences":[{"apiVersion":"apps/v1","kind":"ReplicaSet","name":"cluster-autoscaler-operator-5f49d774cd","uid":"3b719b58-abeb-4a7e-a33a-abec50e3e86e","controller":true,"blockOwnerDeletion":true}]},"spec":{"volumes":[{"name":"cert","secret":{"secretName":"cluster-autoscaler-operator-cert","items":[{"key":"tls.crt","path":"tls.crt"},{"key":"tls.key","path":"tls.key"}],"defaultMode":420}},{"name":"auth-proxy-config","configMap":{"name":"kube-rbac-proxy-cluster-autoscaler-operator","defaultMode":420}},{"name":"kube-api-access-2pqsl","projected":{"sources":[{"serviceAccountToken":{"expirationSeconds":3607,"path":"token"}},{"configMap":{"name":"kube-root-ca.crt","items":[{"key":"ca.crt","path":"ca.crt"}]}},{"downwardAPI":{"items":[{"path":"namespace","fieldRef":{"apiVersion":"v1","fieldPath":"metadata.namespace"}}]}},{"configMap":{"name":"openshift-service-ca.crt","items":[{"key":"service-ca.crt","path":"service-ca.crt"}]}}],"defaultMode":420}}],"containers":[{"name":"kube-rbac-proxy","image":"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69ffd8f8dcceedc2d6eb306cea33f8beabc1be1308cd5f4ee8b9a8e3eab9843","args":["--secure-listen-address=0.0.0.0:9192","--upstream=http://127.0.0.1:9191/","--tls-cert-file=/etc/tls/private/tls.crt","--tls-private-key-file=/etc/tls/private/tls.key","--config-file=/etc/kube-rbac-proxy/config-file.yaml","--tls-cipher-suites=TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305,TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305","--logtostderr=true","--v=3"],"ports":[{"name":"metrics","containerPort":9192,"protocol":"TCP"}],"resources":{"requests":{"cpu":"10m","memory":"20Mi"}},"volumeMounts":[{"name":"auth-proxy-config","readOnly":true,"mountPath":"/etc/kube-rbac-proxy"},{"name":"cert","readOnly":true,"mountPath":"/etc/tls/private"},{"name":"kube-api-access-2pqsl","readOnly":true,"mountPath":"/var/run/secrets/kubernetes.io/serviceaccount"}],"terminationMessagePath":"/dev/termination-log","terminationMessagePolicy":"FallbackToLogsOnError","imagePullPolicy":"IfNotPresent","securityContext":{"capabilities":{"drop":["ALL"]},"runAsUser":1000500000,"runAsNonRoot":true,"allowPrivilegeEscalation":false}},{"name":"cluster-autoscaler-operator","image":"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:72bbe2c638872937108f647950ab8ad35c0428ca8ecc6a39a8314aace7d95078","command":["cluster-autoscaler-operator"],"args":["-alsologtostderr"],"ports":[{"containerPort":8443,"protocol":"TCP"}],"env":[{"name":"RELEASE_VERSION","value":"4.18.29"},{"name":"WATCH_NAMESPACE","valueFrom":{"fieldRef":{"apiVersion":"v1","fieldPath":"metadata.namespace"}}},{"name":"CLUSTER_AUTOSCALER_NAMESPACE","valueFrom":{"fieldRef":{"apiVersion":"v1","fieldPath":"metadata.namespace"}}},{"name":"LEADER_ELECTION_NAMESPACE","valueFrom":{"fieldRef":{"apiVersion":"v1","fieldPath":"metadata.namespace"}}},{"name":"CLUSTER_AUTOSCALER_IMAGE","value":"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:6127d09ff17b35440f5116598731952167eab9b30ad2218efd25fbb3fd0d7586"},{"name":"WEBHOOKS_CERT_DIR","value":"/etc/cluster-autoscaler-operator/tls"},{"name":"WEBHOOKS_PORT","value":"8443"},{"name":"METRICS_PORT","value":"9191"}],"resources":{"requests":{"cpu":"20m","memory":"50Mi"}},"volumeMounts":[{"name":"cert","readOnly":true,"mountPath":"/etc/cluster-autoscaler-operator/tls"},{"name":"kube-api-access-2pqsl","readOnly":true,"mountPath":"/var/run/secrets/kubernetes.io/serviceaccount"}],"terminationMessagePath":"/dev/termination-log","terminationMessagePolicy":"FallbackToLogsOnError","imagePullPolicy":"IfNotPresent","securityContext":{"capabilities":{"drop":["ALL"]},"runAsUser":1000500000,"runAsNonRoot":true,"allowPrivilegeEscalation":false}}],"restartPolicy":"Always","terminationGracePeriodSeconds":30,"dnsPolicy":"ClusterFirst","nodeSelector":{"node-role.kubernetes.io/master":""},"serviceAccountName":"cluster-autoscaler-operator","serviceAccount":"cluster-autoscaler-operator","nodeName":"master-0","securityContext":{"seLinuxOptions":{"level":"s0:c22,c19"},"fsGroup":1000500000,"seccompProfile":{"type":"RuntimeDefault"}},"imagePullSecrets":[{"name":"cluster-autoscaler-operator-dockercfg-zs454"}],"schedulerName":"default-scheduler","tolerations":[{"key":"node-role.kubernetes.io/master","operator":"Exists","effect":"NoSchedule"},{"key":"node.kubernetes.io/not-ready","operator":"Exists","effect":"NoExecute","tolerationSeconds":300},{"key":"node.kubernetes.io/unreachable","operator":"Exists","effect":"NoExecute","tolerationSeconds":300},{"key":"node.kubernetes.io/memory-pressure","operator":"Exists","effect":"NoSchedule"}],"priorityClassName":"system-node-critical","priority":2000001000,"enableServiceLinks":true,"preemptionPolicy":"PreemptLowerPriority"},"status":{"phase":"Running","conditions":[{"type":"PodReadyToStartContainers","status":"True","lastProbeTime":null,"lastTransitionTime":"2025-12-04T11:39:22Z"},{"type":"Initialized","status":"True","lastProbeTime":null,"lastTransitionTime":"2025-12-04T11:38:43Z"},{"type":"Ready","status":"True","lastProbeTime":null,"lastTransitionTime":"2025-12-04T11:56:05Z"},{"type":"ContainersReady","status":"True","lastProbeTime":null,"lastTransitionTime":"2025-12-04T11:56:05Z"},{"type":"PodScheduled","status":"True","lastProbeTime":null,"lastTransitionTime":"2025-12-04T11:38:43Z"}],"hostIP":"192.168.32.10","hostIPs":[{"ip":"192.168.32.10"}],"podIP":"10.128.0.59","podIPs":[{"ip":"10.128.0.59"}],"startTime":"2025-12-04T11:38:43Z","containerStatuses":[{"name":"cluster-autoscaler-operator","state":{"running":{"startedAt":"2025-12-04T11:56:04Z"}},"lastState":{"terminated":{"exitCode":255,"reason":"Error","message":" in the time allotted, but may still be processing the request (get leases.coordination.k8s.io cluster-autoscaler-operator-leader)\nE1204 11:53:04.405651 1 status.go:311] status reporting failed: the server was unable to return a response in the time allotted, but may still be processing the request (get clusteroperators.config.openshift.io cluster-autoscaler)\nE1204 11:54:04.407610 1 status.go:426] failed to get dependency machine-api status: the server was unable to return a response in the time allotted, but may still be processing the request (get clusteroperators.config.openshift.io machine-api)\nW1204 11:54:04.407685 1 status.go:271] Operator status degraded: error checking machine-api status: the server was unable to return a response in the time allotted, but may still be processing the request (get clusteroperators.config.openshift.io machine-api)\nE1204 11:54:12.578608 1 leaderelection.go:429] Failed to update lock optimitically: Timeout: request did not complete within requested timeout - context deadline exceeded, falling back to slow path\nE1204 11:55:04.409885 1 status.go:311] status reporting failed: the server was unable to return a response in the time allotted, but may still be processing the request (get clusteroperators.config.openshift.io cluster-autoscaler)\nE1204 11:55:04.563080 1 leaderelection.go:436] error retrieving resource lock openshift-machine-api/cluster-autoscaler-operator-leader: Get \"https://172.30.0.1:443/apis/coordination.k8s.io/v1/namespaces/openshift-machine-api/leases/cluster-autoscaler-operator-leader\": context deadline exceeded\nI1204 11:55:04.563166 1 leaderelection.go:297] failed to renew lease openshift-machine-api/cluster-autoscaler-operator-leader: timed out waiting for the condition\nE1204 11:55:38.566293 1 leaderelection.go:322] Failed to release lock: Timeout: request did not complete within requested timeout - context deadline exceeded\nF1204 11:55:38.566379 1 main.go:43] Failed to start operator: leader election lost\n","startedAt":"2025-12-04T11:46:04Z","finishedAt":"2025-12-04T11:55:38Z","containerID":"cri-o://f06d1731c63d1c3b95fb1ded6be8b346f6f343eb1c39fb1931d998db924768a0"}},"ready":true,"restartCount":3,"image":"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:72bbe2c638872937108f647950ab8ad35c0428ca8ecc6a39a8314aace7d95078","imageID":"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:72bbe2c638872937108f647950ab8ad35c0428ca8ecc6a39a8314aace7d95078","containerID":"cri-o://29db6007fa122142272c2e747034efd7c5978859754eee3760f27dbc521cf520","started":true,"volumeMounts":[{"name":"cert","mountPath":"/etc/cluster-autoscaler-operator/tls","readOnly":true,"recursiveReadOnly":"Disabled"},{"name":"kube-api-access-2pqsl","mountPath":"/var/run/secrets/kubernetes.io/serviceaccount","readOnly":true,"recursiveReadOnly":"Disabled"}]},{"name":"kube-rbac-proxy","state":{"running":{"startedAt":"2025-12-04T11:38:44Z"}},"lastState":{},"ready":true,"restartCount":0,"image":"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69ffd8f8dcceedc2d6eb306cea33f8beabc1be1308cd5f4ee8b9a8e3eab9843","imageID":"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69ffd8f8dcceedc2d6eb306cea33f8beabc1be1308cd5f4ee8b9a8e3eab9843","containerID":"cri-o://8351b3c33e5381d9f8f52c5cbb85fe67b5c8d17b81f220227d907d879cdef1ea","started":true,"volumeMounts":[{"name":"auth-proxy-config","mountPath":"/etc/kube-rbac-proxy","readOnly":true,"recursiveReadOnly":"Disabled"},{"name":"cert","mountPath":"/etc/tls/private","readOnly":true,"recursiveReadOnly":"Disabled"},{"name":"kube-api-access-2pqsl","mountPath":"/var/run/secrets/kubernetes.io/serviceaccount","readOnly":true,"recursiveReadOnly":"Disabled"}]}],"qosClass":"Burstable"}}kube-storage-version-migrator-operator-b9c5dfc78-dcxkw.json0000640000000000000000000001631715114274061036077 0ustar0000000000000000config/pod/openshift-kube-storage-version-migrator-operator{"metadata":{"name":"kube-storage-version-migrator-operator-b9c5dfc78-dcxkw","generateName":"kube-storage-version-migrator-operator-b9c5dfc78-","namespace":"openshift-kube-storage-version-migrator-operator","uid":"fd9f8671-8066-4990-b45d-8b619aa5d9ec","resourceVersion":"9875","creationTimestamp":"2025-12-04T11:35:50Z","labels":{"app":"kube-storage-version-migrator-operator","pod-template-hash":"b9c5dfc78"},"annotations":{"k8s.ovn.org/pod-networks":"{\"default\":{\"ip_addresses\":[\"10.128.0.26/23\"],\"mac_address\":\"0a:58:0a:80:00:1a\",\"gateway_ips\":[\"10.128.0.1\"],\"routes\":[{\"dest\":\"10.128.0.0/16\",\"nextHop\":\"10.128.0.1\"},{\"dest\":\"172.30.0.0/16\",\"nextHop\":\"10.128.0.1\"},{\"dest\":\"169.254.0.5/32\",\"nextHop\":\"10.128.0.1\"},{\"dest\":\"100.64.0.0/16\",\"nextHop\":\"10.128.0.1\"}],\"ip_address\":\"10.128.0.26/23\",\"gateway_ip\":\"10.128.0.1\",\"role\":\"primary\"}}","k8s.v1.cni.cncf.io/network-status":"[{\n \"name\": \"ovn-kubernetes\",\n \"interface\": \"eth0\",\n \"ips\": [\n \"10.128.0.26\"\n ],\n \"mac\": \"0a:58:0a:80:00:1a\",\n \"default\": true,\n \"dns\": {}\n}]","openshift.io/required-scc":"nonroot-v2","openshift.io/scc":"nonroot-v2","seccomp.security.alpha.kubernetes.io/pod":"runtime/default"},"ownerReferences":[{"apiVersion":"apps/v1","kind":"ReplicaSet","name":"kube-storage-version-migrator-operator-b9c5dfc78","uid":"b23931bb-9b31-4b3a-8a82-0af7fe03b041","controller":true,"blockOwnerDeletion":true}]},"spec":{"volumes":[{"name":"serving-cert","secret":{"secretName":"serving-cert","defaultMode":420,"optional":true}},{"name":"config","configMap":{"name":"config","defaultMode":420}},{"name":"kube-api-access-llscb","projected":{"sources":[{"serviceAccountToken":{"expirationSeconds":3607,"path":"token"}},{"configMap":{"name":"kube-root-ca.crt","items":[{"key":"ca.crt","path":"ca.crt"}]}},{"downwardAPI":{"items":[{"path":"namespace","fieldRef":{"apiVersion":"v1","fieldPath":"metadata.namespace"}}]}},{"configMap":{"name":"openshift-service-ca.crt","items":[{"key":"service-ca.crt","path":"service-ca.crt"}]}}],"defaultMode":420}}],"containers":[{"name":"kube-storage-version-migrator-operator","image":"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:75d996f6147edb88c09fd1a052099de66638590d7d03a735006244bc9e19f898","command":["cluster-kube-storage-version-migrator-operator","start"],"args":["--config=/var/run/configmaps/config/config.yaml"],"ports":[{"name":"metrics","containerPort":8443,"protocol":"TCP"}],"env":[{"name":"IMAGE","value":"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e438b814f8e16f00b3fc4b69991af80eee79ae111d2a707f34aa64b2ccbb6eb"},{"name":"OPERATOR_IMAGE","value":"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:75d996f6147edb88c09fd1a052099de66638590d7d03a735006244bc9e19f898"},{"name":"OPERATOR_IMAGE_VERSION","value":"4.18.29"},{"name":"OPERAND_IMAGE_VERSION","value":"4.18.29"}],"resources":{"requests":{"cpu":"10m","memory":"50Mi"}},"volumeMounts":[{"name":"config","mountPath":"/var/run/configmaps/config"},{"name":"serving-cert","mountPath":"/var/run/secrets/serving-cert"},{"name":"kube-api-access-llscb","readOnly":true,"mountPath":"/var/run/secrets/kubernetes.io/serviceaccount"}],"terminationMessagePath":"/dev/termination-log","terminationMessagePolicy":"FallbackToLogsOnError","imagePullPolicy":"IfNotPresent","securityContext":{"capabilities":{"drop":["ALL"]},"runAsUser":1001,"allowPrivilegeEscalation":false}}],"restartPolicy":"Always","terminationGracePeriodSeconds":30,"dnsPolicy":"ClusterFirst","nodeSelector":{"node-role.kubernetes.io/master":""},"serviceAccountName":"kube-storage-version-migrator-operator","serviceAccount":"kube-storage-version-migrator-operator","nodeName":"master-0","securityContext":{"seLinuxOptions":{"level":"s0:c19,c14"},"runAsNonRoot":true,"seccompProfile":{"type":"RuntimeDefault"}},"schedulerName":"default-scheduler","tolerations":[{"key":"node-role.kubernetes.io/master","operator":"Exists","effect":"NoSchedule"},{"key":"node.kubernetes.io/unreachable","operator":"Exists","effect":"NoExecute","tolerationSeconds":120},{"key":"node.kubernetes.io/not-ready","operator":"Exists","effect":"NoExecute","tolerationSeconds":120},{"key":"node.kubernetes.io/memory-pressure","operator":"Exists","effect":"NoSchedule"}],"priorityClassName":"system-cluster-critical","priority":2000000000,"enableServiceLinks":true,"preemptionPolicy":"PreemptLowerPriority"},"status":{"phase":"Running","conditions":[{"type":"PodReadyToStartContainers","status":"True","lastProbeTime":null,"lastTransitionTime":"2025-12-04T11:37:43Z"},{"type":"Initialized","status":"True","lastProbeTime":null,"lastTransitionTime":"2025-12-04T11:37:30Z"},{"type":"Ready","status":"True","lastProbeTime":null,"lastTransitionTime":"2025-12-04T11:45:10Z"},{"type":"ContainersReady","status":"True","lastProbeTime":null,"lastTransitionTime":"2025-12-04T11:45:10Z"},{"type":"PodScheduled","status":"True","lastProbeTime":null,"lastTransitionTime":"2025-12-04T11:37:30Z"}],"hostIP":"192.168.32.10","hostIPs":[{"ip":"192.168.32.10"}],"podIP":"10.128.0.26","podIPs":[{"ip":"10.128.0.26"}],"startTime":"2025-12-04T11:37:30Z","containerStatuses":[{"name":"kube-storage-version-migrator-operator","state":{"running":{"startedAt":"2025-12-04T11:44:28Z"}},"lastState":{"terminated":{"exitCode":255,"reason":"Error","message":"I1204 11:42:12.313315 1 cmd.go:253] Using service-serving-cert provided certificates\nI1204 11:42:12.313470 1 leaderelection.go:121] The leader election gives 4 retries and allows for 30s of clock skew. The kube-apiserver downtime tolerance is 78s. Worst non-graceful lease acquisition is 2m43s. Worst graceful lease acquisition is {26s}.\nI1204 11:42:12.313974 1 observer_polling.go:159] Starting file observer\nW1204 11:43:12.321395 1 builder.go:272] unable to get owner reference (falling back to namespace): the server was unable to return a response in the time allotted, but may still be processing the request (get pods)\nI1204 11:43:12.321885 1 builder.go:304] openshift-kube-storage-version-migrator-operator version 4.18.0-202511181540.p2.g59ba356.assembly.stream.el9-59ba356-59ba356f50ea3128905ffdb7137f868aa0588bab\nF1204 11:43:42.639863 1 cmd.go:182] failed checking apiserver connectivity: Get \"https://172.30.0.1:443/apis/coordination.k8s.io/v1/namespaces/openshift-kube-storage-version-migrator-operator/leases/openshift-kube-storage-version-migrator-operator-lock\": context deadline exceeded\n","startedAt":"2025-12-04T11:42:12Z","finishedAt":"2025-12-04T11:43:42Z","containerID":"cri-o://313a5a3e45bfc56240a41a6b2d15c82ca825cd210b3e33c0eb74e4e915c86d54"}},"ready":true,"restartCount":3,"image":"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:75d996f6147edb88c09fd1a052099de66638590d7d03a735006244bc9e19f898","imageID":"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:75d996f6147edb88c09fd1a052099de66638590d7d03a735006244bc9e19f898","containerID":"cri-o://1cffe972f64278af41c2f8763a71bbb64f7f719300efa47c1cb1f0a4d8df9393","started":true,"volumeMounts":[{"name":"config","mountPath":"/var/run/configmaps/config"},{"name":"serving-cert","mountPath":"/var/run/secrets/serving-cert"},{"name":"kube-api-access-llscb","mountPath":"/var/run/secrets/kubernetes.io/serviceaccount","readOnly":true,"recursiveReadOnly":"Disabled"}]}],"qosClass":"Burstable"}}config/pod/openshift-kube-scheduler/openshift-kube-scheduler-master-0.json0000640000000000000000000003700015114274061025407 0ustar0000000000000000{"metadata":{"name":"openshift-kube-scheduler-master-0","namespace":"openshift-kube-scheduler","uid":"b1f518f0-9b56-41aa-b487-3cfaeff48585","resourceVersion":"11922","creationTimestamp":"2025-12-04T11:45:52Z","labels":{"app":"openshift-kube-scheduler","revision":"4","scheduler":"true"},"annotations":{"kubectl.kubernetes.io/default-container":"kube-scheduler","kubernetes.io/config.hash":"bdd20c68f1ccfd1240fb273af0638d01","kubernetes.io/config.mirror":"bdd20c68f1ccfd1240fb273af0638d01","kubernetes.io/config.seen":"2025-12-04T11:45:52.512303140Z","kubernetes.io/config.source":"file","target.workload.openshift.io/management":"{\"effect\": \"PreferredDuringScheduling\"}"},"ownerReferences":[{"apiVersion":"v1","kind":"Node","name":"master-0","uid":"9618c86a-8a87-4c00-8dd2-9747684d64b3","controller":true}]},"spec":{"volumes":[{"name":"resource-dir","hostPath":{"path":"/etc/kubernetes/static-pod-resources/kube-scheduler-pod-4","type":""}},{"name":"cert-dir","hostPath":{"path":"/etc/kubernetes/static-pod-resources/kube-scheduler-certs","type":""}}],"initContainers":[{"name":"wait-for-host-port","image":"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:6d5891cdd7dcf7c9081de8b364b4c96446b7f946f7880fbae291a4592a198264","command":["/usr/bin/timeout","30","/bin/bash","-c"],"args":["echo -n \"Waiting for port :10259 to be released.\"\nwhile [ -n \"$(ss -Htan '( sport = 10259 )')\" ]; do\n echo -n \".\"\n sleep 1\ndone\n"],"resources":{"requests":{"cpu":"15m","memory":"50Mi"}},"terminationMessagePath":"/dev/termination-log","terminationMessagePolicy":"FallbackToLogsOnError","imagePullPolicy":"IfNotPresent"}],"containers":[{"name":"kube-scheduler","image":"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:6d5891cdd7dcf7c9081de8b364b4c96446b7f946f7880fbae291a4592a198264","command":["hyperkube","kube-scheduler"],"args":["--config=/etc/kubernetes/static-pod-resources/configmaps/config/config.yaml","--cert-dir=/var/run/kubernetes","--authentication-kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/scheduler-kubeconfig/kubeconfig","--authorization-kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/scheduler-kubeconfig/kubeconfig","--feature-gates=AWSClusterHostedDNS=false,AWSEFSDriverVolumeMetrics=true,AdditionalRoutingCapabilities=false,AdminNetworkPolicy=true,AlibabaPlatform=true,AutomatedEtcdBackup=false,AzureWorkloadIdentity=true,BareMetalLoadBalancer=true,BootcNodeManagement=false,BuildCSIVolumes=true,CSIDriverSharedResource=false,ChunkSizeMiB=true,CloudDualStackNodeIPs=true,ClusterAPIInstall=false,ClusterAPIInstallIBMCloud=false,ClusterMonitoringConfig=false,ConsolePluginContentSecurityPolicy=false,DNSNameResolver=false,DisableKubeletCloudCredentialProviders=true,DynamicResourceAllocation=false,EtcdBackendQuota=false,EventedPLEG=false,Example=false,ExternalOIDC=false,ExternalOIDCWithUIDAndExtraClaimMappings=false,GCPClusterHostedDNS=false,GCPLabelsTags=true,GatewayAPI=false,HardwareSpeed=true,ImageStreamImportMode=false,IngressControllerDynamicConfigurationManager=false,IngressControllerLBSubnetsAWS=true,InsightsConfig=false,InsightsConfigAPI=false,InsightsOnDemandDataGather=false,InsightsRuntimeExtractor=false,KMSv1=true,MachineAPIMigration=false,MachineAPIOperatorDisableMachineHealthCheckController=false,MachineAPIProviderOpenStack=false,MachineConfigNodes=false,ManagedBootImages=true,ManagedBootImagesAWS=true,MaxUnavailableStatefulSet=false,MetricsCollectionProfiles=false,MinimumKubeletVersion=false,MixedCPUsAllocation=false,MultiArchInstallAWS=true,MultiArchInstallAzure=false,MultiArchInstallGCP=true,NetworkDiagnosticsConfig=true,NetworkLiveMigration=true,NetworkSegmentation=true,NewOLM=true,NodeDisruptionPolicy=true,NodeSwap=false,NutanixMultiSubnets=false,OVNObservability=false,OnClusterBuild=true,OpenShiftPodSecurityAdmission=false,PersistentIPsForVirtualization=true,PinnedImages=false,PlatformOperators=false,PrivateHostedZoneAWS=true,ProcMountType=false,RouteAdvertisements=false,RouteExternalCertificate=false,ServiceAccountTokenNodeBinding=false,SetEIPForNLBIngressController=true,SignatureStores=false,SigstoreImageVerification=false,StreamingCollectionEncodingToJSON=false,StreamingCollectionEncodingToProtobuf=false,TranslateStreamCloseWebsocketRequests=false,UpgradeStatus=false,UserNamespacesPodSecurityStandards=false,UserNamespacesSupport=false,VSphereControlPlaneMachineSet=true,VSphereDriverConfiguration=true,VSphereMultiNetworks=false,VSphereMultiVCenters=true,VSphereStaticIPs=true,ValidatingAdmissionPolicy=true,VolumeAttributesClass=false,VolumeGroupSnapshot=false","-v=2","--tls-cert-file=/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt","--tls-private-key-file=/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key","--tls-cipher-suites=TLS_AES_128_GCM_SHA256,TLS_AES_256_GCM_SHA384,TLS_CHACHA20_POLY1305_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305_SHA256,TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305_SHA256","--tls-min-version=VersionTLS12"],"ports":[{"hostPort":10259,"containerPort":10259,"protocol":"TCP"}],"resources":{"requests":{"cpu":"15m","memory":"50Mi"}},"volumeMounts":[{"name":"resource-dir","mountPath":"/etc/kubernetes/static-pod-resources"},{"name":"cert-dir","mountPath":"/etc/kubernetes/static-pod-certs"}],"livenessProbe":{"httpGet":{"path":"healthz","port":10259,"scheme":"HTTPS"},"initialDelaySeconds":45,"timeoutSeconds":1,"periodSeconds":10,"successThreshold":1,"failureThreshold":3},"readinessProbe":{"httpGet":{"path":"healthz","port":10259,"scheme":"HTTPS"},"initialDelaySeconds":45,"timeoutSeconds":1,"periodSeconds":10,"successThreshold":1,"failureThreshold":3},"terminationMessagePath":"/dev/termination-log","terminationMessagePolicy":"FallbackToLogsOnError","imagePullPolicy":"IfNotPresent"},{"name":"kube-scheduler-cert-syncer","image":"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f042fa25014f3d37f3ea967d21f361d2a11833ae18f2c750318101b25d2497ce","command":["cluster-kube-scheduler-operator","cert-syncer"],"args":["--kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/kube-scheduler-cert-syncer-kubeconfig/kubeconfig","--namespace=$(POD_NAMESPACE)","--destination-dir=/etc/kubernetes/static-pod-certs"],"env":[{"name":"POD_NAME","valueFrom":{"fieldRef":{"apiVersion":"v1","fieldPath":"metadata.name"}}},{"name":"POD_NAMESPACE","valueFrom":{"fieldRef":{"apiVersion":"v1","fieldPath":"metadata.namespace"}}}],"resources":{"requests":{"cpu":"5m","memory":"50Mi"}},"volumeMounts":[{"name":"resource-dir","mountPath":"/etc/kubernetes/static-pod-resources"},{"name":"cert-dir","mountPath":"/etc/kubernetes/static-pod-certs"}],"terminationMessagePath":"/dev/termination-log","terminationMessagePolicy":"FallbackToLogsOnError","imagePullPolicy":"IfNotPresent"},{"name":"kube-scheduler-recovery-controller","image":"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f042fa25014f3d37f3ea967d21f361d2a11833ae18f2c750318101b25d2497ce","command":["/bin/bash","-euxo","pipefail","-c"],"args":["timeout 3m /bin/bash -exuo pipefail -c 'while [ -n \"$(ss -Htanop \\( sport = 11443 \\))\" ]; do sleep 1; done'\n\nexec cluster-kube-scheduler-operator cert-recovery-controller --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/kube-scheduler-cert-syncer-kubeconfig/kubeconfig --namespace=${POD_NAMESPACE} --listen=0.0.0.0:11443 -v=2\n"],"env":[{"name":"POD_NAMESPACE","valueFrom":{"fieldRef":{"apiVersion":"v1","fieldPath":"metadata.namespace"}}}],"resources":{"requests":{"cpu":"5m","memory":"50Mi"}},"volumeMounts":[{"name":"resource-dir","mountPath":"/etc/kubernetes/static-pod-resources"},{"name":"cert-dir","mountPath":"/etc/kubernetes/static-pod-certs"}],"terminationMessagePath":"/dev/termination-log","terminationMessagePolicy":"FallbackToLogsOnError","imagePullPolicy":"IfNotPresent"}],"restartPolicy":"Always","terminationGracePeriodSeconds":30,"dnsPolicy":"ClusterFirst","nodeName":"master-0","hostNetwork":true,"securityContext":{},"schedulerName":"default-scheduler","tolerations":[{"operator":"Exists"}],"priorityClassName":"system-node-critical","priority":2000001000,"enableServiceLinks":true,"preemptionPolicy":"PreemptLowerPriority"},"status":{"phase":"Running","conditions":[{"type":"PodReadyToStartContainers","status":"True","lastProbeTime":null,"lastTransitionTime":"2025-12-04T11:59:18Z"},{"type":"Initialized","status":"True","lastProbeTime":null,"lastTransitionTime":"2025-12-04T11:59:18Z"},{"type":"Ready","status":"True","lastProbeTime":null,"lastTransitionTime":"2025-12-04T11:59:22Z"},{"type":"ContainersReady","status":"True","lastProbeTime":null,"lastTransitionTime":"2025-12-04T11:59:22Z"},{"type":"PodScheduled","status":"True","lastProbeTime":null,"lastTransitionTime":"2025-12-04T11:59:18Z"}],"hostIP":"192.168.32.10","hostIPs":[{"ip":"192.168.32.10"}],"podIP":"192.168.32.10","podIPs":[{"ip":"192.168.32.10"}],"startTime":"2025-12-04T11:59:18Z","initContainerStatuses":[{"name":"wait-for-host-port","state":{"terminated":{"exitCode":0,"reason":"Completed","startedAt":"2025-12-04T11:45:53Z","finishedAt":"2025-12-04T11:45:53Z","containerID":"cri-o://f22a84f6fb348e29fba57d1a09fc02c20059513a38fe36937800076d569de9eb"}},"lastState":{},"ready":true,"restartCount":0,"image":"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:6d5891cdd7dcf7c9081de8b364b4c96446b7f946f7880fbae291a4592a198264","imageID":"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:6d5891cdd7dcf7c9081de8b364b4c96446b7f946f7880fbae291a4592a198264","containerID":"cri-o://f22a84f6fb348e29fba57d1a09fc02c20059513a38fe36937800076d569de9eb","started":false}],"containerStatuses":[{"name":"kube-scheduler","state":{"running":{"startedAt":"2025-12-04T11:53:29Z"}},"lastState":{"terminated":{"exitCode":1,"reason":"Error","message":"broadcaster.go:270] \"Server rejected event (will not retry!)\" err=\"Timeout: request did not complete within requested timeout - context deadline exceeded\" event=\"\u0026Event{ObjectMeta:{network-check-source-85d8db45d4-5bjlq.187e00979363e288 openshift-network-diagnostics 0 0001-01-01 00:00:00 +0000 UTC \u003cnil\u003e \u003cnil\u003e map[] map[] [] [] []},EventTime:2025-12-04 11:45:54.988670694 +0000 UTC m=+0.949415529,Series:\u0026EventSeries{Count:2,LastObservedTime:2025-12-04 11:50:54.99646345 +0000 UTC m=+300.957208335,},ReportingController:default-scheduler,ReportingInstance:default-scheduler-master-0,Action:Scheduling,Reason:FailedScheduling,Regarding:{Pod openshift-network-diagnostics network-check-source-85d8db45d4-5bjlq 0372b35f-9414-43a9-b739-509eaababbb1 v1 3108 },Related:nil,Note:0/1 nodes are available: 1 node(s) had untolerated taint {node-role.kubernetes.io/master: }. preemption: 0/1 nodes are available: 1 Preemption is not helpful for scheduling.,Type:Warning,DeprecatedSource:{ },DeprecatedFirstTimestamp:0001-01-01 00:00:00 +0000 UTC,DeprecatedLastTimestamp:0001-01-01 00:00:00 +0000 UTC,DeprecatedCount:0,}\"\nE1204 11:51:41.498643 1 leaderelection.go:429] Failed to update lock optimitically: Timeout: request did not complete within requested timeout - context deadline exceeded, falling back to slow path\nE1204 11:52:34.999680 1 leaderelection.go:436] error retrieving resource lock openshift-kube-scheduler/kube-scheduler: Get \"https://api-int.sno.openstack.lab:6443/apis/coordination.k8s.io/v1/namespaces/openshift-kube-scheduler/leases/kube-scheduler?timeout=53.5s\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)\nI1204 11:52:54.496570 1 leaderelection.go:297] failed to renew lease openshift-kube-scheduler/kube-scheduler: timed out waiting for the condition\nE1204 11:53:28.500096 1 leaderelection.go:322] Failed to release lock: Timeout: request did not complete within requested timeout - context deadline exceeded\nE1204 11:53:28.500149 1 server.go:309] \"Leaderelection lost\"\n","startedAt":"2025-12-04T11:45:54Z","finishedAt":"2025-12-04T11:53:28Z","containerID":"cri-o://e892e4a03cb775cc984461c05f66ebb315e6e3256740bfe4765f9efa5a82a35d"}},"ready":true,"restartCount":1,"image":"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:6d5891cdd7dcf7c9081de8b364b4c96446b7f946f7880fbae291a4592a198264","imageID":"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:6d5891cdd7dcf7c9081de8b364b4c96446b7f946f7880fbae291a4592a198264","containerID":"cri-o://e2713a2b9adf8ac2cd64c11a8cc47639892df9751d3f6828f4bd8b7f6c344a80","started":true},{"name":"kube-scheduler-cert-syncer","state":{"running":{"startedAt":"2025-12-04T11:55:55Z"}},"lastState":{"terminated":{"exitCode":1,"reason":"Error","message":"ourceVersion=0\": tls: failed to verify certificate: x509: certificate is valid for kubernetes, kubernetes.default, kubernetes.default.svc, kubernetes.default.svc.cluster.local, openshift, openshift.default, openshift.default.svc, openshift.default.svc.cluster.local, not localhost-recovery\nE1204 11:55:07.170231 1 reflector.go:158] \"Unhandled Error\" err=\"k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243: Failed to watch *v1.Secret: failed to list *v1.Secret: Get \\\"https://localhost:6443/api/v1/namespaces/openshift-kube-scheduler/secrets?limit=500\u0026resourceVersion=0\\\": tls: failed to verify certificate: x509: certificate is valid for kubernetes, kubernetes.default, kubernetes.default.svc, kubernetes.default.svc.cluster.local, openshift, openshift.default, openshift.default.svc, openshift.default.svc.cluster.local, not localhost-recovery\"\nW1204 11:55:33.111623 1 reflector.go:561] k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243: failed to list *v1.ConfigMap: Get \"https://localhost:6443/api/v1/namespaces/openshift-kube-scheduler/configmaps?limit=500\u0026resourceVersion=0\": tls: failed to verify certificate: x509: certificate is valid for kubernetes, kubernetes.default, kubernetes.default.svc, kubernetes.default.svc.cluster.local, openshift, openshift.default, openshift.default.svc, openshift.default.svc.cluster.local, not localhost-recovery\nE1204 11:55:33.111691 1 reflector.go:158] \"Unhandled Error\" err=\"k8s.io/client-go@v0.31.1/tools/cache/reflector.go:243: Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: Get \\\"https://localhost:6443/api/v1/namespaces/openshift-kube-scheduler/configmaps?limit=500\u0026resourceVersion=0\\\": tls: failed to verify certificate: x509: certificate is valid for kubernetes, kubernetes.default, kubernetes.default.svc, kubernetes.default.svc.cluster.local, openshift, openshift.default, openshift.default.svc, openshift.default.svc.cluster.local, not localhost-recovery\"\nF1204 11:55:54.361620 1 base_controller.go:105] unable to sync caches for CertSyncController\n","startedAt":"2025-12-04T11:45:54Z","finishedAt":"2025-12-04T11:55:54Z","containerID":"cri-o://83e55c475e3dc5500479af8907cda3d13a24a31c8ee5c2278bdd8d400f5e7292"}},"ready":true,"restartCount":1,"image":"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f042fa25014f3d37f3ea967d21f361d2a11833ae18f2c750318101b25d2497ce","imageID":"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f042fa25014f3d37f3ea967d21f361d2a11833ae18f2c750318101b25d2497ce","containerID":"cri-o://9e6711a29e80dd1a67221ee7d3cb99dca0b39d86d535ebc251bf63200ef5ce12","started":true},{"name":"kube-scheduler-recovery-controller","state":{"running":{"startedAt":"2025-12-04T11:45:54Z"}},"lastState":{},"ready":true,"restartCount":0,"image":"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f042fa25014f3d37f3ea967d21f361d2a11833ae18f2c750318101b25d2497ce","imageID":"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f042fa25014f3d37f3ea967d21f361d2a11833ae18f2c750318101b25d2497ce","containerID":"cri-o://1dd3e9bec426604298540a62160aa04995ba35e82eac44c41bf666b47786e5c9","started":true}],"qosClass":"Burstable"}}config/pod/openshift-kube-apiserver-operator/kube-apiserver-operator-765d9ff747-gr68k.json0000640000000000000000000001600715114274061030007 0ustar0000000000000000{"metadata":{"name":"kube-apiserver-operator-765d9ff747-gr68k","generateName":"kube-apiserver-operator-765d9ff747-","namespace":"openshift-kube-apiserver-operator","uid":"ccf81538-4d5a-4c26-8e54-fe39331a625d","resourceVersion":"9818","creationTimestamp":"2025-12-04T11:35:49Z","labels":{"app":"kube-apiserver-operator","pod-template-hash":"765d9ff747"},"annotations":{"k8s.ovn.org/pod-networks":"{\"default\":{\"ip_addresses\":[\"10.128.0.25/23\"],\"mac_address\":\"0a:58:0a:80:00:19\",\"gateway_ips\":[\"10.128.0.1\"],\"routes\":[{\"dest\":\"10.128.0.0/16\",\"nextHop\":\"10.128.0.1\"},{\"dest\":\"172.30.0.0/16\",\"nextHop\":\"10.128.0.1\"},{\"dest\":\"169.254.0.5/32\",\"nextHop\":\"10.128.0.1\"},{\"dest\":\"100.64.0.0/16\",\"nextHop\":\"10.128.0.1\"}],\"ip_address\":\"10.128.0.25/23\",\"gateway_ip\":\"10.128.0.1\",\"role\":\"primary\"}}","k8s.v1.cni.cncf.io/network-status":"[{\n \"name\": \"ovn-kubernetes\",\n \"interface\": \"eth0\",\n \"ips\": [\n \"10.128.0.25\"\n ],\n \"mac\": \"0a:58:0a:80:00:19\",\n \"default\": true,\n \"dns\": {}\n}]"},"ownerReferences":[{"apiVersion":"apps/v1","kind":"ReplicaSet","name":"kube-apiserver-operator-765d9ff747","uid":"ef9e2c47-de1f-447c-8bd3-3bf0fd5ec6cd","controller":true,"blockOwnerDeletion":true}]},"spec":{"volumes":[{"name":"serving-cert","secret":{"secretName":"kube-apiserver-operator-serving-cert","defaultMode":420,"optional":true}},{"name":"config","configMap":{"name":"kube-apiserver-operator-config","defaultMode":420}},{"name":"kube-api-access","projected":{"sources":[{"serviceAccountToken":{"expirationSeconds":3600,"path":"token"}},{"configMap":{"name":"kube-root-ca.crt","items":[{"key":"ca.crt","path":"ca.crt"}]}},{"downwardAPI":{"items":[{"path":"namespace","fieldRef":{"apiVersion":"v1","fieldPath":"metadata.namespace"}}]}}],"defaultMode":420}}],"containers":[{"name":"kube-apiserver-operator","image":"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:0f43c31aa3359159d4557dad3cfaf812d8ce44db9cb9ae970e06d3479070b660","command":["cluster-kube-apiserver-operator","operator"],"args":["--config=/var/run/configmaps/config/config.yaml"],"ports":[{"name":"metrics","containerPort":8443,"protocol":"TCP"}],"env":[{"name":"IMAGE","value":"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:6d5891cdd7dcf7c9081de8b364b4c96446b7f946f7880fbae291a4592a198264"},{"name":"OPERATOR_IMAGE","value":"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:0f43c31aa3359159d4557dad3cfaf812d8ce44db9cb9ae970e06d3479070b660"},{"name":"OPERAND_IMAGE_VERSION","value":"1.31.13"},{"name":"OPERATOR_IMAGE_VERSION","value":"4.18.29"},{"name":"POD_NAME","valueFrom":{"fieldRef":{"apiVersion":"v1","fieldPath":"metadata.name"}}}],"resources":{"requests":{"cpu":"10m","memory":"50Mi"}},"volumeMounts":[{"name":"config","mountPath":"/var/run/configmaps/config"},{"name":"serving-cert","mountPath":"/var/run/secrets/serving-cert"},{"name":"kube-api-access","readOnly":true,"mountPath":"/var/run/secrets/kubernetes.io/serviceaccount"}],"terminationMessagePath":"/dev/termination-log","terminationMessagePolicy":"FallbackToLogsOnError","imagePullPolicy":"IfNotPresent","securityContext":{"capabilities":{"drop":["ALL"]},"allowPrivilegeEscalation":false}}],"restartPolicy":"Always","terminationGracePeriodSeconds":30,"dnsPolicy":"ClusterFirst","nodeSelector":{"node-role.kubernetes.io/master":""},"serviceAccountName":"kube-apiserver-operator","serviceAccount":"kube-apiserver-operator","automountServiceAccountToken":false,"nodeName":"master-0","securityContext":{"runAsUser":65534,"runAsNonRoot":true,"seccompProfile":{"type":"RuntimeDefault"}},"schedulerName":"default-scheduler","tolerations":[{"key":"node-role.kubernetes.io/master","operator":"Exists","effect":"NoSchedule"},{"key":"node-role.kubernetes.io/control-plane","operator":"Exists","effect":"NoExecute"},{"key":"node.kubernetes.io/unreachable","operator":"Exists","effect":"NoExecute","tolerationSeconds":120},{"key":"node.kubernetes.io/not-ready","operator":"Exists","effect":"NoExecute","tolerationSeconds":120},{"key":"node.kubernetes.io/memory-pressure","operator":"Exists","effect":"NoSchedule"}],"priorityClassName":"system-cluster-critical","priority":2000000000,"enableServiceLinks":true,"preemptionPolicy":"PreemptLowerPriority"},"status":{"phase":"Running","conditions":[{"type":"PodReadyToStartContainers","status":"True","lastProbeTime":null,"lastTransitionTime":"2025-12-04T11:37:35Z"},{"type":"Initialized","status":"True","lastProbeTime":null,"lastTransitionTime":"2025-12-04T11:37:30Z"},{"type":"Ready","status":"True","lastProbeTime":null,"lastTransitionTime":"2025-12-04T11:44:27Z"},{"type":"ContainersReady","status":"True","lastProbeTime":null,"lastTransitionTime":"2025-12-04T11:44:27Z"},{"type":"PodScheduled","status":"True","lastProbeTime":null,"lastTransitionTime":"2025-12-04T11:37:30Z"}],"hostIP":"192.168.32.10","hostIPs":[{"ip":"192.168.32.10"}],"podIP":"10.128.0.25","podIPs":[{"ip":"10.128.0.25"}],"startTime":"2025-12-04T11:37:30Z","containerStatuses":[{"name":"kube-apiserver-operator","state":{"running":{"startedAt":"2025-12-04T11:43:20Z"}},"lastState":{"terminated":{"exitCode":255,"reason":"Error","message":"I1204 11:41:38.514927 1 cmd.go:253] Using service-serving-cert provided certificates\nI1204 11:41:38.515147 1 leaderelection.go:121] The leader election gives 4 retries and allows for 30s of clock skew. The kube-apiserver downtime tolerance is 78s. Worst non-graceful lease acquisition is 2m43s. Worst graceful lease acquisition is {26s}.\nI1204 11:41:38.515648 1 observer_polling.go:159] Starting file observer\nW1204 11:42:38.523455 1 builder.go:272] unable to get owner reference (falling back to namespace): the server was unable to return a response in the time allotted, but may still be processing the request (get pods kube-apiserver-operator-765d9ff747-gr68k)\nI1204 11:42:38.523604 1 builder.go:304] kube-apiserver-operator version 4.18.0-202511181540.p2.gbac8a7c.assembly.stream.el9-bac8a7c-bac8a7c99c2e8562a05b784c8fe4f8ecd91eecdf\nF1204 11:43:08.978788 1 cmd.go:182] failed checking apiserver connectivity: Get \"https://172.30.0.1:443/apis/coordination.k8s.io/v1/namespaces/openshift-kube-apiserver-operator/leases/kube-apiserver-operator-lock\": context deadline exceeded\n","startedAt":"2025-12-04T11:41:38Z","finishedAt":"2025-12-04T11:43:08Z","containerID":"cri-o://6740e091e7513192b7c4fc2e37bc11133fd10dfdaf1d0800e4f08d8808fea4fc"}},"ready":true,"restartCount":2,"image":"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:0f43c31aa3359159d4557dad3cfaf812d8ce44db9cb9ae970e06d3479070b660","imageID":"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:0f43c31aa3359159d4557dad3cfaf812d8ce44db9cb9ae970e06d3479070b660","containerID":"cri-o://fa82cd61790ebe17dba45d17ac4fc09f5a6705740362c96127dc7edfc6054c63","started":true,"volumeMounts":[{"name":"config","mountPath":"/var/run/configmaps/config"},{"name":"serving-cert","mountPath":"/var/run/secrets/serving-cert"},{"name":"kube-api-access","mountPath":"/var/run/secrets/kubernetes.io/serviceaccount","readOnly":true,"recursiveReadOnly":"Disabled"}]}],"qosClass":"Burstable"}}config/pod/openshift-cluster-storage-operator/csi-snapshot-controller-6b958b6f94-w74zr.json0000640000000000000000000001402115114274061030247 0ustar0000000000000000{"metadata":{"name":"csi-snapshot-controller-6b958b6f94-w74zr","generateName":"csi-snapshot-controller-6b958b6f94-","namespace":"openshift-cluster-storage-operator","uid":"38fc8086-a00c-4a2a-8a0e-c57e9d9d0103","resourceVersion":"11430","creationTimestamp":"2025-12-04T11:37:44Z","labels":{"app":"csi-snapshot-controller","pod-template-hash":"6b958b6f94"},"annotations":{"k8s.ovn.org/pod-networks":"{\"default\":{\"ip_addresses\":[\"10.128.0.27/23\"],\"mac_address\":\"0a:58:0a:80:00:1b\",\"gateway_ips\":[\"10.128.0.1\"],\"routes\":[{\"dest\":\"10.128.0.0/16\",\"nextHop\":\"10.128.0.1\"},{\"dest\":\"172.30.0.0/16\",\"nextHop\":\"10.128.0.1\"},{\"dest\":\"169.254.0.5/32\",\"nextHop\":\"10.128.0.1\"},{\"dest\":\"100.64.0.0/16\",\"nextHop\":\"10.128.0.1\"}],\"ip_address\":\"10.128.0.27/23\",\"gateway_ip\":\"10.128.0.1\",\"role\":\"primary\"}}","k8s.v1.cni.cncf.io/network-status":"[{\n \"name\": \"ovn-kubernetes\",\n \"interface\": \"eth0\",\n \"ips\": [\n \"10.128.0.27\"\n ],\n \"mac\": \"0a:58:0a:80:00:1b\",\n \"default\": true,\n \"dns\": {}\n}]","openshift.io/required-scc":"restricted-v2","openshift.io/scc":"restricted-v2","seccomp.security.alpha.kubernetes.io/pod":"runtime/default"},"ownerReferences":[{"apiVersion":"apps/v1","kind":"ReplicaSet","name":"csi-snapshot-controller-6b958b6f94","uid":"6200de80-59e5-4601-9da2-92ec2592e107","controller":true,"blockOwnerDeletion":true}]},"spec":{"volumes":[{"name":"kube-api-access-l5r7w","projected":{"sources":[{"serviceAccountToken":{"expirationSeconds":3607,"path":"token"}},{"configMap":{"name":"kube-root-ca.crt","items":[{"key":"ca.crt","path":"ca.crt"}]}},{"downwardAPI":{"items":[{"path":"namespace","fieldRef":{"apiVersion":"v1","fieldPath":"metadata.namespace"}}]}},{"configMap":{"name":"openshift-service-ca.crt","items":[{"key":"service-ca.crt","path":"service-ca.crt"}]}}],"defaultMode":420}}],"containers":[{"name":"snapshot-controller","image":"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d3ce2cbf1032ad0f24f204db73687002fcf302e86ebde3945801c74351b64576","args":["--v=2","--leader-election=true","--leader-election-lease-duration=137s","--leader-election-renew-deadline=107s","--leader-election-retry-period=26s","--leader-election-namespace=openshift-cluster-storage-operator"],"resources":{"requests":{"cpu":"10m","memory":"50Mi"}},"volumeMounts":[{"name":"kube-api-access-l5r7w","readOnly":true,"mountPath":"/var/run/secrets/kubernetes.io/serviceaccount"}],"terminationMessagePath":"/dev/termination-log","terminationMessagePolicy":"FallbackToLogsOnError","imagePullPolicy":"IfNotPresent","securityContext":{"capabilities":{"drop":["ALL"]},"runAsUser":1000170000,"allowPrivilegeEscalation":false}}],"restartPolicy":"Always","terminationGracePeriodSeconds":30,"dnsPolicy":"ClusterFirst","nodeSelector":{"node-role.kubernetes.io/master":""},"serviceAccountName":"csi-snapshot-controller","serviceAccount":"csi-snapshot-controller","nodeName":"master-0","securityContext":{"seLinuxOptions":{"level":"s0:c13,c7"},"runAsNonRoot":true,"fsGroup":1000170000,"seccompProfile":{"type":"RuntimeDefault"}},"affinity":{"podAntiAffinity":{"preferredDuringSchedulingIgnoredDuringExecution":[{"weight":100,"podAffinityTerm":{"labelSelector":{"matchLabels":{"app":"csi-snapshot-controller"}},"topologyKey":"kubernetes.io/hostname"}}]}},"schedulerName":"default-scheduler","tolerations":[{"key":"node.kubernetes.io/unreachable","operator":"Exists","effect":"NoExecute","tolerationSeconds":120},{"key":"node.kubernetes.io/not-ready","operator":"Exists","effect":"NoExecute","tolerationSeconds":120},{"key":"node-role.kubernetes.io/master","operator":"Exists","effect":"NoSchedule"},{"key":"node.kubernetes.io/memory-pressure","operator":"Exists","effect":"NoSchedule"}],"priorityClassName":"system-cluster-critical","priority":2000000000,"enableServiceLinks":true,"preemptionPolicy":"PreemptLowerPriority"},"status":{"phase":"Running","conditions":[{"type":"PodReadyToStartContainers","status":"True","lastProbeTime":null,"lastTransitionTime":"2025-12-04T11:37:52Z"},{"type":"Initialized","status":"True","lastProbeTime":null,"lastTransitionTime":"2025-12-04T11:37:44Z"},{"type":"Ready","status":"True","lastProbeTime":null,"lastTransitionTime":"2025-12-04T11:57:40Z"},{"type":"ContainersReady","status":"True","lastProbeTime":null,"lastTransitionTime":"2025-12-04T11:57:40Z"},{"type":"PodScheduled","status":"True","lastProbeTime":null,"lastTransitionTime":"2025-12-04T11:37:44Z"}],"hostIP":"192.168.32.10","hostIPs":[{"ip":"192.168.32.10"}],"podIP":"10.128.0.27","podIPs":[{"ip":"10.128.0.27"}],"startTime":"2025-12-04T11:37:44Z","containerStatuses":[{"name":"snapshot-controller","state":{"running":{"startedAt":"2025-12-04T11:57:39Z"}},"lastState":{"terminated":{"exitCode":1,"reason":"Error","message":"I1204 11:54:16.007275 1 feature_gate.go:387] feature gates: {map[]}\nI1204 11:54:16.007524 1 main.go:169] Version: 28bdee88181a0d0c59b906003ceb29f73dc8562b\nI1204 11:54:16.009068 1 main.go:220] Start NewCSISnapshotController with kubeconfig [] resyncPeriod [15m0s]\nE1204 11:54:46.012252 1 main.go:98] Failed to list v1 volumesnapshots with error=Get \"https://172.30.0.1:443/apis/snapshot.storage.k8s.io/v1/volumesnapshots?limit=1\": context deadline exceeded\nE1204 11:54:46.012394 1 main.go:246] Exiting due to failure to ensure CRDs exist during startup: context deadline exceeded\n","startedAt":"2025-12-04T11:54:15Z","finishedAt":"2025-12-04T11:54:46Z","containerID":"cri-o://c3d6e3ec1c2075f0d2c2e41480f7486576efc2d0e3c1b8e7a0df0f84b64a6eb3"}},"ready":true,"restartCount":6,"image":"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d3ce2cbf1032ad0f24f204db73687002fcf302e86ebde3945801c74351b64576","imageID":"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d3ce2cbf1032ad0f24f204db73687002fcf302e86ebde3945801c74351b64576","containerID":"cri-o://e6dd3de655999dd73139dd633fcda105be0e68a4757cbd984035ddeac7e8ca3f","started":true,"volumeMounts":[{"name":"kube-api-access-l5r7w","mountPath":"/var/run/secrets/kubernetes.io/serviceaccount","readOnly":true,"recursiveReadOnly":"Disabled"}]}],"qosClass":"Burstable"}}config/pod/openshift-cluster-node-tuning-operator/cluster-node-tuning-operator-85cff47f46-4gv5j.json0000640000000000000000000002061315114274061032021 0ustar0000000000000000{"metadata":{"name":"cluster-node-tuning-operator-85cff47f46-4gv5j","generateName":"cluster-node-tuning-operator-85cff47f46-","namespace":"openshift-cluster-node-tuning-operator","uid":"6f15dcdb-c968-429c-9d1d-63c2becad337","resourceVersion":"11274","creationTimestamp":"2025-12-04T11:35:48Z","labels":{"name":"cluster-node-tuning-operator","pod-template-hash":"85cff47f46"},"annotations":{"k8s.ovn.org/pod-networks":"{\"default\":{\"ip_addresses\":[\"10.128.0.7/23\"],\"mac_address\":\"0a:58:0a:80:00:07\",\"gateway_ips\":[\"10.128.0.1\"],\"routes\":[{\"dest\":\"10.128.0.0/16\",\"nextHop\":\"10.128.0.1\"},{\"dest\":\"172.30.0.0/16\",\"nextHop\":\"10.128.0.1\"},{\"dest\":\"169.254.0.5/32\",\"nextHop\":\"10.128.0.1\"},{\"dest\":\"100.64.0.0/16\",\"nextHop\":\"10.128.0.1\"}],\"ip_address\":\"10.128.0.7/23\",\"gateway_ip\":\"10.128.0.1\",\"role\":\"primary\"}}","k8s.v1.cni.cncf.io/network-status":"[{\n \"name\": \"ovn-kubernetes\",\n \"interface\": \"eth0\",\n \"ips\": [\n \"10.128.0.7\"\n ],\n \"mac\": \"0a:58:0a:80:00:07\",\n \"default\": true,\n \"dns\": {}\n}]","openshift.io/required-scc":"anyuid","openshift.io/scc":"anyuid"},"ownerReferences":[{"apiVersion":"apps/v1","kind":"ReplicaSet","name":"cluster-node-tuning-operator-85cff47f46","uid":"d42b02f0-bb11-4968-8fb1-c14f4d73c02d","controller":true,"blockOwnerDeletion":true}]},"spec":{"volumes":[{"name":"node-tuning-operator-tls","secret":{"secretName":"node-tuning-operator-tls","defaultMode":420}},{"name":"apiservice-cert","secret":{"secretName":"performance-addon-operator-webhook-cert","items":[{"key":"tls.crt","path":"apiserver.crt"},{"key":"tls.key","path":"apiserver.key"}],"defaultMode":420}},{"name":"trusted-ca","configMap":{"name":"trusted-ca","items":[{"key":"ca-bundle.crt","path":"tls-ca-bundle.pem"}],"defaultMode":420,"optional":true}},{"name":"kube-api-access-kkjzz","projected":{"sources":[{"serviceAccountToken":{"expirationSeconds":3607,"path":"token"}},{"configMap":{"name":"kube-root-ca.crt","items":[{"key":"ca.crt","path":"ca.crt"}]}},{"downwardAPI":{"items":[{"path":"namespace","fieldRef":{"apiVersion":"v1","fieldPath":"metadata.namespace"}}]}},{"configMap":{"name":"openshift-service-ca.crt","items":[{"key":"service-ca.crt","path":"service-ca.crt"}]}}],"defaultMode":420}}],"containers":[{"name":"cluster-node-tuning-operator","image":"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5451aa441e5b8d8689c032405d410c8049a849ef2edf77e5b6a5ce2838c6569b","command":["cluster-node-tuning-operator"],"args":["-v=0"],"ports":[{"name":"metrics","containerPort":60000,"protocol":"TCP"}],"env":[{"name":"RELEASE_VERSION","value":"4.18.29"},{"name":"WATCH_NAMESPACE","valueFrom":{"fieldRef":{"apiVersion":"v1","fieldPath":"metadata.namespace"}}},{"name":"POD_NAME","valueFrom":{"fieldRef":{"apiVersion":"v1","fieldPath":"metadata.name"}}},{"name":"RESYNC_PERIOD","value":"600"},{"name":"CLUSTER_NODE_TUNED_IMAGE","value":"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5451aa441e5b8d8689c032405d410c8049a849ef2edf77e5b6a5ce2838c6569b"}],"resources":{"requests":{"cpu":"10m","memory":"20Mi"}},"volumeMounts":[{"name":"node-tuning-operator-tls","mountPath":"/etc/secrets"},{"name":"trusted-ca","mountPath":"/var/run/configmaps/trusted-ca/"},{"name":"apiservice-cert","mountPath":"/apiserver.local.config/certificates"},{"name":"kube-api-access-kkjzz","readOnly":true,"mountPath":"/var/run/secrets/kubernetes.io/serviceaccount"}],"terminationMessagePath":"/dev/termination-log","terminationMessagePolicy":"FallbackToLogsOnError","imagePullPolicy":"IfNotPresent","securityContext":{"capabilities":{"drop":["MKNOD"]},"readOnlyRootFilesystem":true}}],"restartPolicy":"Always","terminationGracePeriodSeconds":30,"dnsPolicy":"ClusterFirst","nodeSelector":{"node-role.kubernetes.io/master":""},"serviceAccountName":"cluster-node-tuning-operator","serviceAccount":"cluster-node-tuning-operator","nodeName":"master-0","securityContext":{"seLinuxOptions":{"level":"s0:c14,c9"},"runAsUser":499,"runAsNonRoot":true},"schedulerName":"default-scheduler","tolerations":[{"key":"node-role.kubernetes.io/master","operator":"Exists","effect":"NoSchedule"},{"key":"node.kubernetes.io/unreachable","operator":"Exists","effect":"NoExecute","tolerationSeconds":120},{"key":"node.kubernetes.io/not-ready","operator":"Exists","effect":"NoExecute","tolerationSeconds":120},{"key":"node.kubernetes.io/memory-pressure","operator":"Exists","effect":"NoSchedule"}],"priorityClassName":"system-cluster-critical","priority":2000000000,"enableServiceLinks":true,"preemptionPolicy":"PreemptLowerPriority"},"status":{"phase":"Running","conditions":[{"type":"PodReadyToStartContainers","status":"True","lastProbeTime":null,"lastTransitionTime":"2025-12-04T11:38:02Z"},{"type":"Initialized","status":"True","lastProbeTime":null,"lastTransitionTime":"2025-12-04T11:37:30Z"},{"type":"Ready","status":"True","lastProbeTime":null,"lastTransitionTime":"2025-12-04T11:55:48Z"},{"type":"ContainersReady","status":"True","lastProbeTime":null,"lastTransitionTime":"2025-12-04T11:55:48Z"},{"type":"PodScheduled","status":"True","lastProbeTime":null,"lastTransitionTime":"2025-12-04T11:37:30Z"}],"hostIP":"192.168.32.10","hostIPs":[{"ip":"192.168.32.10"}],"podIP":"10.128.0.7","podIPs":[{"ip":"10.128.0.7"}],"startTime":"2025-12-04T11:37:30Z","containerStatuses":[{"name":"cluster-node-tuning-operator","state":{"running":{"startedAt":"2025-12-04T11:55:47Z"}},"lastState":{"terminated":{"exitCode":1,"reason":"Error","message":",\"msg\":\"Starting EventSource\",\"controller\":\"performanceprofile\",\"controllerGroup\":\"performance.openshift.io\",\"controllerKind\":\"PerformanceProfile\",\"source\":\"kind source: *v1.Profile\"}\n{\"level\":\"info\",\"ts\":\"2025-12-04T11:50:12Z\",\"msg\":\"Starting Controller\",\"controller\":\"performanceprofile\",\"controllerGroup\":\"performance.openshift.io\",\"controllerKind\":\"PerformanceProfile\"}\nI1204 11:50:13.072102 1 controller.go:1443] started events processor/controller\nI1204 11:50:13.077498 1 server.go:104] starting metrics server\n{\"level\":\"info\",\"ts\":\"2025-12-04T11:50:13Z\",\"msg\":\"Starting workers\",\"controller\":\"performanceprofile\",\"controllerGroup\":\"performance.openshift.io\",\"controllerKind\":\"PerformanceProfile\",\"worker count\":1}\nE1204 11:51:46.988242 1 leaderelection.go:429] Failed to update lock optimitically: Timeout: request did not complete within requested timeout - context deadline exceeded, falling back to slow path\nE1204 11:52:46.990311 1 leaderelection.go:436] error retrieving resource lock openshift-cluster-node-tuning-operator/node-tuning-operator-lock: the server was unable to return a response in the time allotted, but may still be processing the request (get leases.coordination.k8s.io node-tuning-operator-lock)\nE1204 11:54:20.995859 1 leaderelection.go:429] Failed to update lock optimitically: Timeout: request did not complete within requested timeout - context deadline exceeded, falling back to slow path\nE1204 11:55:12.985029 1 leaderelection.go:436] error retrieving resource lock openshift-cluster-node-tuning-operator/node-tuning-operator-lock: Get \"https://172.30.0.1:443/apis/coordination.k8s.io/v1/namespaces/openshift-cluster-node-tuning-operator/leases/node-tuning-operator-lock\": context deadline exceeded\nI1204 11:55:12.985112 1 leaderelection.go:297] failed to renew lease openshift-cluster-node-tuning-operator/node-tuning-operator-lock: timed out waiting for the condition\nF1204 11:55:46.783690 1 main.go:242] manager exited with non-zero code: leader election lost\n","startedAt":"2025-12-04T11:45:02Z","finishedAt":"2025-12-04T11:55:46Z","containerID":"cri-o://a2ccae65671ed8e1587834a54d81391f1b1881f39ee38aa28b193fa6cecc39da"}},"ready":true,"restartCount":2,"image":"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5451aa441e5b8d8689c032405d410c8049a849ef2edf77e5b6a5ce2838c6569b","imageID":"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5451aa441e5b8d8689c032405d410c8049a849ef2edf77e5b6a5ce2838c6569b","containerID":"cri-o://be438243594f3bf861e4019351de704cf0f44ae021f8193424a9d6d9b91f168c","started":true,"volumeMounts":[{"name":"node-tuning-operator-tls","mountPath":"/etc/secrets"},{"name":"trusted-ca","mountPath":"/var/run/configmaps/trusted-ca/"},{"name":"apiservice-cert","mountPath":"/apiserver.local.config/certificates"},{"name":"kube-api-access-kkjzz","mountPath":"/var/run/secrets/kubernetes.io/serviceaccount","readOnly":true,"recursiveReadOnly":"Disabled"}]}],"qosClass":"Burstable"}}config/pod/openshift-cluster-machine-approver/machine-approver-f797d8546-qvgbq.json0000640000000000000000000002133415114274061026645 0ustar0000000000000000{"metadata":{"name":"machine-approver-f797d8546-qvgbq","generateName":"machine-approver-f797d8546-","namespace":"openshift-cluster-machine-approver","uid":"c835660f-5a79-4679-b9be-8543ca6ef5a1","resourceVersion":"11254","creationTimestamp":"2025-12-04T11:38:37Z","labels":{"app":"machine-approver","pod-template-hash":"f797d8546"},"ownerReferences":[{"apiVersion":"apps/v1","kind":"ReplicaSet","name":"machine-approver-f797d8546","uid":"cfea205c-f5ed-457c-a47e-ddf5351f36c2","controller":true,"blockOwnerDeletion":true}]},"spec":{"volumes":[{"name":"auth-proxy-config","configMap":{"name":"kube-rbac-proxy","defaultMode":420}},{"name":"machine-approver-tls","secret":{"secretName":"machine-approver-tls","defaultMode":420}},{"name":"config","configMap":{"name":"machine-approver-config","defaultMode":440,"optional":true}},{"name":"kube-api-access-nxbts","projected":{"sources":[{"serviceAccountToken":{"expirationSeconds":3607,"path":"token"}},{"configMap":{"name":"kube-root-ca.crt","items":[{"key":"ca.crt","path":"ca.crt"}]}},{"downwardAPI":{"items":[{"path":"namespace","fieldRef":{"apiVersion":"v1","fieldPath":"metadata.namespace"}}]}},{"configMap":{"name":"openshift-service-ca.crt","items":[{"key":"service-ca.crt","path":"service-ca.crt"}]}}],"defaultMode":420}}],"containers":[{"name":"kube-rbac-proxy","image":"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69ffd8f8dcceedc2d6eb306cea33f8beabc1be1308cd5f4ee8b9a8e3eab9843","args":["--secure-listen-address=0.0.0.0:9192","--upstream=http://127.0.0.1:9191/","--tls-cert-file=/etc/tls/private/tls.crt","--tls-private-key-file=/etc/tls/private/tls.key","--config-file=/etc/kube-rbac-proxy/config-file.yaml","--tls-cipher-suites=TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305,TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305","--logtostderr=true","--v=3"],"ports":[{"name":"https","hostPort":9192,"containerPort":9192,"protocol":"TCP"}],"resources":{"requests":{"cpu":"10m","memory":"20Mi"}},"volumeMounts":[{"name":"auth-proxy-config","mountPath":"/etc/kube-rbac-proxy"},{"name":"machine-approver-tls","mountPath":"/etc/tls/private"},{"name":"kube-api-access-nxbts","readOnly":true,"mountPath":"/var/run/secrets/kubernetes.io/serviceaccount"}],"terminationMessagePath":"/dev/termination-log","terminationMessagePolicy":"FallbackToLogsOnError","imagePullPolicy":"IfNotPresent"},{"name":"machine-approver-controller","image":"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8cc27777e72233024fe84ee1faa168aec715a0b24912a3ce70715ddccba328df","command":["/usr/bin/machine-approver"],"args":["--config=/var/run/configmaps/config/config.yaml","-v=2","--logtostderr","--leader-elect=true","--leader-elect-lease-duration=137s","--leader-elect-renew-deadline=107s","--leader-elect-retry-period=26s","--leader-elect-resource-namespace=openshift-cluster-machine-approver","--api-group-version=machine.openshift.io/v1beta1","--max-concurrent-reconciles=10"],"env":[{"name":"RELEASE_VERSION","value":"4.18.29"}],"resources":{"requests":{"cpu":"10m","memory":"50Mi"}},"volumeMounts":[{"name":"config","mountPath":"/var/run/configmaps/config"},{"name":"kube-api-access-nxbts","readOnly":true,"mountPath":"/var/run/secrets/kubernetes.io/serviceaccount"}],"terminationMessagePath":"/dev/termination-log","terminationMessagePolicy":"FallbackToLogsOnError","imagePullPolicy":"IfNotPresent"}],"restartPolicy":"Always","terminationGracePeriodSeconds":30,"dnsPolicy":"ClusterFirst","nodeSelector":{"node-role.kubernetes.io/master":""},"serviceAccountName":"machine-approver-sa","serviceAccount":"machine-approver-sa","nodeName":"master-0","hostNetwork":true,"securityContext":{},"schedulerName":"default-scheduler","tolerations":[{"key":"node-role.kubernetes.io/master","operator":"Exists","effect":"NoSchedule"},{"key":"node.kubernetes.io/unreachable","operator":"Exists","effect":"NoExecute","tolerationSeconds":120},{"key":"node.kubernetes.io/not-ready","operator":"Exists","effect":"NoExecute","tolerationSeconds":120},{"key":"node.kubernetes.io/memory-pressure","operator":"Exists","effect":"NoSchedule"}],"priorityClassName":"system-cluster-critical","priority":2000000000,"enableServiceLinks":true,"preemptionPolicy":"PreemptLowerPriority"},"status":{"phase":"Running","conditions":[{"type":"PodReadyToStartContainers","status":"True","lastProbeTime":null,"lastTransitionTime":"2025-12-04T11:39:22Z"},{"type":"Initialized","status":"True","lastProbeTime":null,"lastTransitionTime":"2025-12-04T11:38:37Z"},{"type":"Ready","status":"True","lastProbeTime":null,"lastTransitionTime":"2025-12-04T11:53:41Z"},{"type":"ContainersReady","status":"True","lastProbeTime":null,"lastTransitionTime":"2025-12-04T11:53:41Z"},{"type":"PodScheduled","status":"True","lastProbeTime":null,"lastTransitionTime":"2025-12-04T11:38:37Z"}],"hostIP":"192.168.32.10","hostIPs":[{"ip":"192.168.32.10"}],"podIP":"192.168.32.10","podIPs":[{"ip":"192.168.32.10"}],"startTime":"2025-12-04T11:38:37Z","containerStatuses":[{"name":"kube-rbac-proxy","state":{"running":{"startedAt":"2025-12-04T11:38:40Z"}},"lastState":{},"ready":true,"restartCount":0,"image":"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69ffd8f8dcceedc2d6eb306cea33f8beabc1be1308cd5f4ee8b9a8e3eab9843","imageID":"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69ffd8f8dcceedc2d6eb306cea33f8beabc1be1308cd5f4ee8b9a8e3eab9843","containerID":"cri-o://77e87b65d2d4db1f1231cca5b248512d4a871ef2353103a6010989d0ff1fc4d1","started":true,"volumeMounts":[{"name":"auth-proxy-config","mountPath":"/etc/kube-rbac-proxy"},{"name":"machine-approver-tls","mountPath":"/etc/tls/private"},{"name":"kube-api-access-nxbts","mountPath":"/var/run/secrets/kubernetes.io/serviceaccount","readOnly":true,"recursiveReadOnly":"Disabled"}]},{"name":"machine-approver-controller","state":{"running":{"startedAt":"2025-12-04T11:53:40Z"}},"lastState":{"terminated":{"exitCode":255,"reason":"Error","message":"e leader\" logger=\"events\" type=\"Normal\" object={\"kind\":\"Lease\",\"namespace\":\"openshift-cluster-machine-approver\",\"name\":\"cluster-machine-approver-leader\",\"uid\":\"8f87cf69-076c-46d0-ad05-318e769aa401\",\"apiVersion\":\"coordination.k8s.io/v1\",\"resourceVersion\":\"9612\"} reason=\"LeaderElection\"\nI1204 11:44:21.641659 1 reflector.go:368] Caches populated for *v1.ClusterOperator from github.com/openshift/cluster-machine-approver/status.go:99\nI1204 11:44:21.773018 1 reflector.go:368] Caches populated for *v1.ConfigMap from sigs.k8s.io/controller-runtime/pkg/cache/internal/informers.go:106\nI1204 11:44:21.842109 1 controller.go:217] \"Starting workers\" controller=\"certificatesigningrequest\" controllerGroup=\"certificates.k8s.io\" controllerKind=\"CertificateSigningRequest\" worker count=10\nI1204 11:45:21.740237 1 status.go:163] Error syncing status trigger: failed to get clusterOperator \"machine-approver\": the server was unable to return a response in the time allotted, but may still be processing the request (get clusteroperators.config.openshift.io machine-approver)\nE1204 11:51:52.901241 1 leaderelection.go:429] Failed to update lock optimitically: Timeout: request did not complete within requested timeout - context deadline exceeded, falling back to slow path\nE1204 11:52:52.903643 1 leaderelection.go:436] error retrieving resource lock openshift-cluster-machine-approver/cluster-machine-approver-leader: the server was unable to return a response in the time allotted, but may still be processing the request (get leases.coordination.k8s.io cluster-machine-approver-leader)\nI1204 11:53:05.893753 1 leaderelection.go:297] failed to renew lease openshift-cluster-machine-approver/cluster-machine-approver-leader: timed out waiting for the condition\nE1204 11:53:39.901477 1 leaderelection.go:322] Failed to release lock: Timeout: request did not complete within requested timeout - context deadline exceeded\nF1204 11:53:39.901568 1 main.go:244] unable to run the manager: leader election lost\n","startedAt":"2025-12-04T11:39:21Z","finishedAt":"2025-12-04T11:53:39Z","containerID":"cri-o://749b44ec9e8d4f00f1e655c70a503d3b1ce1085e0a7eecca2e01f8fe90a6e0c2"}},"ready":true,"restartCount":1,"image":"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8cc27777e72233024fe84ee1faa168aec715a0b24912a3ce70715ddccba328df","imageID":"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8cc27777e72233024fe84ee1faa168aec715a0b24912a3ce70715ddccba328df","containerID":"cri-o://699a63b2d7941ed4336c1635f023cff1f0486ae2fd4b4873876780d919314e24","started":true,"volumeMounts":[{"name":"config","mountPath":"/var/run/configmaps/config"},{"name":"kube-api-access-nxbts","mountPath":"/var/run/secrets/kubernetes.io/serviceaccount","readOnly":true,"recursiveReadOnly":"Disabled"}]}],"qosClass":"Burstable"}}cluster-cloud-controller-manager-operator-74f484689c-jmfn2.json0000640000000000000000000003372115114274061035330 0ustar0000000000000000config/pod/openshift-cloud-controller-manager-operator{"metadata":{"name":"cluster-cloud-controller-manager-operator-74f484689c-jmfn2","generateName":"cluster-cloud-controller-manager-operator-74f484689c-","namespace":"openshift-cloud-controller-manager-operator","uid":"f79f39e0-4aa6-4eea-b8f5-7b5ef18e8d48","resourceVersion":"11230","creationTimestamp":"2025-12-04T11:38:45Z","labels":{"k8s-app":"cloud-manager-operator","pod-template-hash":"74f484689c"},"annotations":{"kubectl.kubernetes.io/default-container":"cluster-cloud-controller-manager"},"ownerReferences":[{"apiVersion":"apps/v1","kind":"ReplicaSet","name":"cluster-cloud-controller-manager-operator-74f484689c","uid":"4d9ffb33-03ab-4b04-aa3e-5c4a0d9d4235","controller":true,"blockOwnerDeletion":true}]},"spec":{"volumes":[{"name":"images","configMap":{"name":"cloud-controller-manager-images","defaultMode":420}},{"name":"host-etc-kube","hostPath":{"path":"/etc/kubernetes","type":"Directory"}},{"name":"auth-proxy-config","configMap":{"name":"kube-rbac-proxy","defaultMode":420}},{"name":"cloud-controller-manager-operator-tls","secret":{"secretName":"cloud-controller-manager-operator-tls","defaultMode":420,"optional":true}},{"name":"kube-api-access-6fssw","projected":{"sources":[{"serviceAccountToken":{"expirationSeconds":3607,"path":"token"}},{"configMap":{"name":"kube-root-ca.crt","items":[{"key":"ca.crt","path":"ca.crt"}]}},{"downwardAPI":{"items":[{"path":"namespace","fieldRef":{"apiVersion":"v1","fieldPath":"metadata.namespace"}}]}},{"configMap":{"name":"openshift-service-ca.crt","items":[{"key":"service-ca.crt","path":"service-ca.crt"}]}}],"defaultMode":420}}],"containers":[{"name":"cluster-cloud-controller-manager","image":"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd38b8be3af889b0f97e2df41517c89a11260901432a9a1ee943195bb3a22737","command":["/bin/bash","-c","#!/bin/bash\nset -o allexport\nif [[ -f /etc/kubernetes/apiserver-url.env ]]; then\n source /etc/kubernetes/apiserver-url.env\nelse\n URL_ONLY_KUBECONFIG=/etc/kubernetes/kubeconfig\nfi\nexec /cluster-controller-manager-operator \\\n--leader-elect=true \\\n--leader-elect-lease-duration=137s \\\n--leader-elect-renew-deadline=107s \\\n--leader-elect-retry-period=26s \\\n--leader-elect-resource-namespace=openshift-cloud-controller-manager-operator \\\n\"--images-json=/etc/cloud-controller-manager-config/images.json\" \\\n--metrics-bind-address=127.0.0.1:9257 \\\n--health-addr=127.0.0.1:9259\n"],"ports":[{"name":"metrics","hostPort":9257,"containerPort":9257,"protocol":"TCP"},{"name":"healthz","hostPort":9259,"containerPort":9259,"protocol":"TCP"}],"env":[{"name":"RELEASE_VERSION","value":"4.18.29"}],"resources":{"requests":{"cpu":"10m","memory":"50Mi"}},"volumeMounts":[{"name":"images","mountPath":"/etc/cloud-controller-manager-config/"},{"name":"host-etc-kube","readOnly":true,"mountPath":"/etc/kubernetes"},{"name":"kube-api-access-6fssw","readOnly":true,"mountPath":"/var/run/secrets/kubernetes.io/serviceaccount"}],"terminationMessagePath":"/dev/termination-log","terminationMessagePolicy":"FallbackToLogsOnError","imagePullPolicy":"IfNotPresent"},{"name":"config-sync-controllers","image":"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd38b8be3af889b0f97e2df41517c89a11260901432a9a1ee943195bb3a22737","command":["/bin/bash","-c","#!/bin/bash\nset -o allexport\nif [[ -f /etc/kubernetes/apiserver-url.env ]]; then\n source /etc/kubernetes/apiserver-url.env\nelse\n URL_ONLY_KUBECONFIG=/etc/kubernetes/kubeconfig\nfi\nexec /config-sync-controllers \\\n--leader-elect=true \\\n--leader-elect-lease-duration=137s \\\n--leader-elect-renew-deadline=107s \\\n--leader-elect-retry-period=26s \\\n--leader-elect-resource-namespace=openshift-cloud-controller-manager-operator \\\n--health-addr=127.0.0.1:9260\n"],"ports":[{"name":"healthz","hostPort":9260,"containerPort":9260,"protocol":"TCP"}],"env":[{"name":"RELEASE_VERSION","value":"4.18.29"}],"resources":{"requests":{"cpu":"10m","memory":"25Mi"}},"volumeMounts":[{"name":"host-etc-kube","readOnly":true,"mountPath":"/etc/kubernetes"},{"name":"kube-api-access-6fssw","readOnly":true,"mountPath":"/var/run/secrets/kubernetes.io/serviceaccount"}],"terminationMessagePath":"/dev/termination-log","terminationMessagePolicy":"FallbackToLogsOnError","imagePullPolicy":"IfNotPresent"},{"name":"kube-rbac-proxy","image":"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69ffd8f8dcceedc2d6eb306cea33f8beabc1be1308cd5f4ee8b9a8e3eab9843","args":["--secure-listen-address=0.0.0.0:9258","--upstream=http://127.0.0.1:9257/","--tls-cert-file=/etc/tls/private/tls.crt","--tls-private-key-file=/etc/tls/private/tls.key","--config-file=/etc/kube-rbac-proxy/config-file.yaml","--tls-cipher-suites=TLS_ECDHE_RSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_ECDSA_WITH_AES_128_GCM_SHA256,TLS_ECDHE_RSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_ECDSA_WITH_AES_256_GCM_SHA384,TLS_ECDHE_RSA_WITH_CHACHA20_POLY1305,TLS_ECDHE_ECDSA_WITH_CHACHA20_POLY1305","--logtostderr=true","--v=3"],"ports":[{"name":"https","hostPort":9258,"containerPort":9258,"protocol":"TCP"}],"resources":{"requests":{"cpu":"10m","memory":"20Mi"}},"volumeMounts":[{"name":"auth-proxy-config","mountPath":"/etc/kube-rbac-proxy"},{"name":"cloud-controller-manager-operator-tls","mountPath":"/etc/tls/private"},{"name":"kube-api-access-6fssw","readOnly":true,"mountPath":"/var/run/secrets/kubernetes.io/serviceaccount"}],"terminationMessagePath":"/dev/termination-log","terminationMessagePolicy":"FallbackToLogsOnError","imagePullPolicy":"IfNotPresent"}],"restartPolicy":"Always","terminationGracePeriodSeconds":30,"dnsPolicy":"ClusterFirst","nodeSelector":{"node-role.kubernetes.io/master":""},"serviceAccountName":"cluster-cloud-controller-manager","serviceAccount":"cluster-cloud-controller-manager","nodeName":"master-0","hostNetwork":true,"securityContext":{},"schedulerName":"default-scheduler","tolerations":[{"key":"node-role.kubernetes.io/master","operator":"Exists","effect":"NoSchedule"},{"key":"node.kubernetes.io/unreachable","operator":"Exists","effect":"NoExecute","tolerationSeconds":120},{"key":"node.kubernetes.io/not-ready","operator":"Exists","effect":"NoExecute","tolerationSeconds":120},{"key":"node.cloudprovider.kubernetes.io/uninitialized","operator":"Exists","effect":"NoSchedule"},{"key":"node.kubernetes.io/not-ready","operator":"Exists","effect":"NoSchedule"},{"key":"node.kubernetes.io/memory-pressure","operator":"Exists","effect":"NoSchedule"}],"priorityClassName":"system-node-critical","priority":2000001000,"enableServiceLinks":true,"preemptionPolicy":"PreemptLowerPriority"},"status":{"phase":"Running","conditions":[{"type":"PodReadyToStartContainers","status":"True","lastProbeTime":null,"lastTransitionTime":"2025-12-04T11:39:25Z"},{"type":"Initialized","status":"True","lastProbeTime":null,"lastTransitionTime":"2025-12-04T11:38:45Z"},{"type":"Ready","status":"True","lastProbeTime":null,"lastTransitionTime":"2025-12-04T11:52:59Z"},{"type":"ContainersReady","status":"True","lastProbeTime":null,"lastTransitionTime":"2025-12-04T11:52:59Z"},{"type":"PodScheduled","status":"True","lastProbeTime":null,"lastTransitionTime":"2025-12-04T11:38:45Z"}],"hostIP":"192.168.32.10","hostIPs":[{"ip":"192.168.32.10"}],"podIP":"192.168.32.10","podIPs":[{"ip":"192.168.32.10"}],"startTime":"2025-12-04T11:38:45Z","containerStatuses":[{"name":"cluster-cloud-controller-manager","state":{"running":{"startedAt":"2025-12-04T11:52:59Z"}},"lastState":{"terminated":{"exitCode":1,"reason":"Error","message":"\"\nI1204 11:44:15.042492 1 controller.go:183] \"Starting Controller\" logger=\"CCMOperator\" controller=\"clusteroperator\" controllerGroup=\"config.openshift.io\" controllerKind=\"ClusterOperator\"\nI1204 11:44:15.153297 1 controller.go:217] \"Starting workers\" logger=\"CCMOperator\" controller=\"clusteroperator\" controllerGroup=\"config.openshift.io\" controllerKind=\"ClusterOperator\" worker count=1\nI1204 11:45:53.472404 1 clusteroperator_controller.go:252] Platform does not require an external cloud provider. Skipping...\nI1204 11:45:53.495840 1 clusteroperator_controller.go:252] Platform does not require an external cloud provider. Skipping...\nI1204 11:50:13.095147 1 clusteroperator_controller.go:252] Platform does not require an external cloud provider. Skipping...\nI1204 11:50:19.100749 1 clusteroperator_controller.go:252] Platform does not require an external cloud provider. Skipping...\nI1204 11:50:22.694026 1 clusteroperator_controller.go:252] Platform does not require an external cloud provider. Skipping...\nE1204 11:51:45.572935 1 leaderelection.go:429] Failed to update lock optimitically: Timeout: request did not complete within requested timeout - context deadline exceeded, falling back to slow path\nE1204 11:52:45.575069 1 leaderelection.go:436] error retrieving resource lock openshift-cloud-controller-manager-operator/cluster-cloud-controller-manager-leader: the server was unable to return a response in the time allotted, but may still be processing the request (get leases.coordination.k8s.io cluster-cloud-controller-manager-leader)\nI1204 11:52:58.569225 1 leaderelection.go:297] failed to renew lease openshift-cloud-controller-manager-operator/cluster-cloud-controller-manager-leader: timed out waiting for the condition\nE1204 11:52:58.569400 1 main.go:227] \"problem running manager\" err=\"leader election lost\" logger=\"CCMOperator.setup\"\nI1204 11:52:58.570026 1 internal.go:538] \"Stopping and waiting for non leader election runnables\" logger=\"CCMOperator\"\n","startedAt":"2025-12-04T11:39:24Z","finishedAt":"2025-12-04T11:52:58Z","containerID":"cri-o://a88d1f0e1d80ac478f7ad73d99091a69eb94f507aee6c58fd48df49c97a44d8d"}},"ready":true,"restartCount":1,"image":"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd38b8be3af889b0f97e2df41517c89a11260901432a9a1ee943195bb3a22737","imageID":"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd38b8be3af889b0f97e2df41517c89a11260901432a9a1ee943195bb3a22737","containerID":"cri-o://e6c2fccf4e6fd6d2ffc424d145ecefb24910bddae93bdeeb548a2d4e20d0f630","started":true,"volumeMounts":[{"name":"images","mountPath":"/etc/cloud-controller-manager-config/"},{"name":"host-etc-kube","mountPath":"/etc/kubernetes","readOnly":true,"recursiveReadOnly":"Disabled"},{"name":"kube-api-access-6fssw","mountPath":"/var/run/secrets/kubernetes.io/serviceaccount","readOnly":true,"recursiveReadOnly":"Disabled"}]},{"name":"config-sync-controllers","state":{"running":{"startedAt":"2025-12-04T11:52:43Z"}},"lastState":{"terminated":{"exitCode":1,"reason":"Error","message":"d4814d\"\nI1204 11:45:53.512160 1 trusted_ca_bundle_controller.go:156] cloud-config was not found: ConfigMap \"cloud-conf\" not found\nI1204 11:45:53.514157 1 cloud_config_sync_controller.go:69] cloud-config sync is not needed, returning early\nI1204 11:45:53.522408 1 cloud_config_sync_controller.go:69] cloud-config sync is not needed, returning early\nI1204 11:45:53.610377 1 cloud_config_sync_controller.go:69] cloud-config sync is not needed, returning early\nI1204 11:45:53.612481 1 trusted_ca_bundle_controller.go:156] cloud-config was not found: ConfigMap \"cloud-conf\" not found\nI1204 11:45:53.617172 1 cloud_config_sync_controller.go:69] cloud-config sync is not needed, returning early\nI1204 11:45:53.623517 1 cloud_config_sync_controller.go:69] cloud-config sync is not needed, returning early\nI1204 11:45:53.629396 1 cloud_config_sync_controller.go:69] cloud-config sync is not needed, returning early\nI1204 11:45:53.634281 1 cloud_config_sync_controller.go:69] cloud-config sync is not needed, returning early\nI1204 11:45:53.639315 1 cloud_config_sync_controller.go:69] cloud-config sync is not needed, returning early\nE1204 11:51:30.351866 1 leaderelection.go:429] Failed to update lock optimitically: Timeout: request did not complete within requested timeout - context deadline exceeded, falling back to slow path\nE1204 11:52:30.353922 1 leaderelection.go:436] error retrieving resource lock openshift-cloud-controller-manager-operator/cluster-cloud-config-sync-leader: the server was unable to return a response in the time allotted, but may still be processing the request (get leases.coordination.k8s.io cluster-cloud-config-sync-leader)\nI1204 11:52:43.349117 1 leaderelection.go:297] failed to renew lease openshift-cloud-controller-manager-operator/cluster-cloud-config-sync-leader: timed out waiting for the condition\nE1204 11:52:43.349249 1 main.go:172] \"problem running manager\" err=\"leader election lost\" logger=\"CCCMOConfigSyncControllers.setup\"\n","startedAt":"2025-12-04T11:39:24Z","finishedAt":"2025-12-04T11:52:43Z","containerID":"cri-o://7169aa966534042a209dbc77d82db08548c2b38ca834a1a1b58302b3c9862901"}},"ready":true,"restartCount":1,"image":"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd38b8be3af889b0f97e2df41517c89a11260901432a9a1ee943195bb3a22737","imageID":"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd38b8be3af889b0f97e2df41517c89a11260901432a9a1ee943195bb3a22737","containerID":"cri-o://522499db1b7f9c87ec01e1460e866cf06b4497c66a47056b8d594ba69f505901","started":true,"volumeMounts":[{"name":"host-etc-kube","mountPath":"/etc/kubernetes","readOnly":true,"recursiveReadOnly":"Disabled"},{"name":"kube-api-access-6fssw","mountPath":"/var/run/secrets/kubernetes.io/serviceaccount","readOnly":true,"recursiveReadOnly":"Disabled"}]},{"name":"kube-rbac-proxy","state":{"running":{"startedAt":"2025-12-04T11:39:25Z"}},"lastState":{},"ready":true,"restartCount":0,"image":"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69ffd8f8dcceedc2d6eb306cea33f8beabc1be1308cd5f4ee8b9a8e3eab9843","imageID":"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69ffd8f8dcceedc2d6eb306cea33f8beabc1be1308cd5f4ee8b9a8e3eab9843","containerID":"cri-o://760657f653245b292cc835a469bd76ffdf70307718fd7139083e006a5e12f961","started":true,"volumeMounts":[{"name":"auth-proxy-config","mountPath":"/etc/kube-rbac-proxy"},{"name":"cloud-controller-manager-operator-tls","mountPath":"/etc/tls/private"},{"name":"kube-api-access-6fssw","mountPath":"/var/run/secrets/kubernetes.io/serviceaccount","readOnly":true,"recursiveReadOnly":"Disabled"}]}],"qosClass":"Burstable"}}config/pod/openshift-catalogd/catalogd-controller-manager-7cc89f4c4c-fd9pv.json0000640000000000000000000002537115114274061026022 0ustar0000000000000000{"metadata":{"name":"catalogd-controller-manager-7cc89f4c4c-fd9pv","generateName":"catalogd-controller-manager-7cc89f4c4c-","namespace":"openshift-catalogd","uid":"ee9a47b1-cdd9-43bc-8115-8851f29106e6","resourceVersion":"11950","creationTimestamp":"2025-12-04T11:38:18Z","labels":{"control-plane":"catalogd-controller-manager","pod-template-hash":"7cc89f4c4c"},"annotations":{"k8s.ovn.org/pod-networks":"{\"default\":{\"ip_addresses\":[\"10.128.0.40/23\"],\"mac_address\":\"0a:58:0a:80:00:28\",\"gateway_ips\":[\"10.128.0.1\"],\"routes\":[{\"dest\":\"10.128.0.0/16\",\"nextHop\":\"10.128.0.1\"},{\"dest\":\"172.30.0.0/16\",\"nextHop\":\"10.128.0.1\"},{\"dest\":\"169.254.0.5/32\",\"nextHop\":\"10.128.0.1\"},{\"dest\":\"100.64.0.0/16\",\"nextHop\":\"10.128.0.1\"}],\"ip_address\":\"10.128.0.40/23\",\"gateway_ip\":\"10.128.0.1\",\"role\":\"primary\"}}","k8s.v1.cni.cncf.io/network-status":"[{\n \"name\": \"ovn-kubernetes\",\n \"interface\": \"eth0\",\n \"ips\": [\n \"10.128.0.40\"\n ],\n \"mac\": \"0a:58:0a:80:00:28\",\n \"default\": true,\n \"dns\": {}\n}]","kubectl.kubernetes.io/default-container":"manager","openshift.io/required-scc":"privileged","openshift.io/scc":"privileged","seccomp.security.alpha.kubernetes.io/pod":"runtime/default"},"ownerReferences":[{"apiVersion":"apps/v1","kind":"ReplicaSet","name":"catalogd-controller-manager-7cc89f4c4c","uid":"836ea146-c1e3-4e6b-b896-8d529fd21273","controller":true,"blockOwnerDeletion":true}]},"spec":{"volumes":[{"name":"cache","emptyDir":{}},{"name":"catalogserver-certs","secret":{"secretName":"catalogserver-cert","items":[{"key":"tls.crt","path":"tls.crt"},{"key":"tls.key","path":"tls.key"}],"defaultMode":420,"optional":false}},{"name":"ca-certs","projected":{"sources":[{"configMap":{"name":"catalogd-trusted-ca-bundle","items":[{"key":"ca-bundle.crt","path":"ca-bundle.crt"}],"optional":false}},{"configMap":{"name":"openshift-service-ca.crt","items":[{"key":"service-ca.crt","path":"service-ca.crt"}],"optional":false}}],"defaultMode":420}},{"name":"etc-containers","hostPath":{"path":"/etc/containers","type":"Directory"}},{"name":"etc-docker","hostPath":{"path":"/etc/docker","type":"Directory"}},{"name":"kube-api-access-rnfmb","projected":{"sources":[{"serviceAccountToken":{"expirationSeconds":3607,"path":"token"}},{"configMap":{"name":"kube-root-ca.crt","items":[{"key":"ca.crt","path":"ca.crt"}]}},{"downwardAPI":{"items":[{"path":"namespace","fieldRef":{"apiVersion":"v1","fieldPath":"metadata.namespace"}}]}},{"configMap":{"name":"openshift-service-ca.crt","items":[{"key":"service-ca.crt","path":"service-ca.crt"}]}}],"defaultMode":420}}],"containers":[{"name":"kube-rbac-proxy","image":"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69ffd8f8dcceedc2d6eb306cea33f8beabc1be1308cd5f4ee8b9a8e3eab9843","args":["--secure-listen-address=0.0.0.0:7443","--http2-disable","--upstream=http://127.0.0.1:8080/","--logtostderr=true","--v=2"],"ports":[{"name":"https","containerPort":7443,"protocol":"TCP"}],"resources":{"requests":{"cpu":"5m","memory":"64Mi"}},"volumeMounts":[{"name":"kube-api-access-rnfmb","readOnly":true,"mountPath":"/var/run/secrets/kubernetes.io/serviceaccount"}],"terminationMessagePath":"/dev/termination-log","terminationMessagePolicy":"FallbackToLogsOnError","imagePullPolicy":"IfNotPresent","securityContext":{"capabilities":{"drop":["ALL"]},"allowPrivilegeEscalation":false}},{"name":"manager","image":"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3f0aa9cd04713acc5c6fea721bd849e1500da8ae945e0b32000887f34d786e0b","command":["./manager"],"args":["--leader-elect","--metrics-bind-address=127.0.0.1:8080","--external-address=catalogd-service.openshift-catalogd.svc","--tls-cert=/var/certs/tls.crt","--tls-key=/var/certs/tls.key","--v=2","--global-pull-secret=openshift-config/pull-secret"],"env":[{"name":"SSL_CERT_DIR","value":"/var/ca-certs"}],"resources":{"requests":{"cpu":"100m","memory":"200Mi"}},"volumeMounts":[{"name":"cache","mountPath":"/var/cache/"},{"name":"catalogserver-certs","mountPath":"/var/certs"},{"name":"ca-certs","readOnly":true,"mountPath":"/var/ca-certs"},{"name":"etc-containers","readOnly":true,"mountPath":"/etc/containers"},{"name":"etc-docker","readOnly":true,"mountPath":"/etc/docker"},{"name":"kube-api-access-rnfmb","readOnly":true,"mountPath":"/var/run/secrets/kubernetes.io/serviceaccount"}],"livenessProbe":{"httpGet":{"path":"/healthz","port":8081,"scheme":"HTTP"},"initialDelaySeconds":15,"timeoutSeconds":1,"periodSeconds":20,"successThreshold":1,"failureThreshold":3},"readinessProbe":{"httpGet":{"path":"/readyz","port":8081,"scheme":"HTTP"},"initialDelaySeconds":5,"timeoutSeconds":1,"periodSeconds":10,"successThreshold":1,"failureThreshold":3},"terminationMessagePath":"/dev/termination-log","terminationMessagePolicy":"FallbackToLogsOnError","imagePullPolicy":"IfNotPresent","securityContext":{"capabilities":{"drop":["ALL"]},"allowPrivilegeEscalation":false}}],"restartPolicy":"Always","terminationGracePeriodSeconds":10,"dnsPolicy":"ClusterFirst","nodeSelector":{"kubernetes.io/os":"linux","node-role.kubernetes.io/master":""},"serviceAccountName":"catalogd-controller-manager","serviceAccount":"catalogd-controller-manager","nodeName":"master-0","securityContext":{"seLinuxOptions":{"type":"spc_t"},"runAsNonRoot":true,"seccompProfile":{"type":"RuntimeDefault"}},"affinity":{"nodeAffinity":{"requiredDuringSchedulingIgnoredDuringExecution":{"nodeSelectorTerms":[{"matchExpressions":[{"key":"kubernetes.io/arch","operator":"In","values":["amd64","arm64","ppc64le","s390x"]},{"key":"kubernetes.io/os","operator":"In","values":["linux"]}]}]}}},"schedulerName":"default-scheduler","tolerations":[{"key":"node-role.kubernetes.io/master","operator":"Exists","effect":"NoSchedule"},{"key":"node.kubernetes.io/unreachable","operator":"Exists","effect":"NoExecute","tolerationSeconds":120},{"key":"node.kubernetes.io/not-ready","operator":"Exists","effect":"NoExecute","tolerationSeconds":120},{"key":"node.kubernetes.io/memory-pressure","operator":"Exists","effect":"NoSchedule"}],"priorityClassName":"system-cluster-critical","priority":2000000000,"enableServiceLinks":true,"preemptionPolicy":"PreemptLowerPriority"},"status":{"phase":"Running","conditions":[{"type":"PodReadyToStartContainers","status":"True","lastProbeTime":null,"lastTransitionTime":"2025-12-04T11:38:21Z"},{"type":"Initialized","status":"True","lastProbeTime":null,"lastTransitionTime":"2025-12-04T11:38:18Z"},{"type":"Ready","status":"True","lastProbeTime":null,"lastTransitionTime":"2025-12-04T11:59:30Z"},{"type":"ContainersReady","status":"True","lastProbeTime":null,"lastTransitionTime":"2025-12-04T11:59:30Z"},{"type":"PodScheduled","status":"True","lastProbeTime":null,"lastTransitionTime":"2025-12-04T11:38:18Z"}],"hostIP":"192.168.32.10","hostIPs":[{"ip":"192.168.32.10"}],"podIP":"10.128.0.40","podIPs":[{"ip":"10.128.0.40"}],"startTime":"2025-12-04T11:38:18Z","containerStatuses":[{"name":"kube-rbac-proxy","state":{"running":{"startedAt":"2025-12-04T11:38:20Z"}},"lastState":{},"ready":true,"restartCount":0,"image":"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69ffd8f8dcceedc2d6eb306cea33f8beabc1be1308cd5f4ee8b9a8e3eab9843","imageID":"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69ffd8f8dcceedc2d6eb306cea33f8beabc1be1308cd5f4ee8b9a8e3eab9843","containerID":"cri-o://12b3881df5a7c036eddc497f15dd545447cd37cf5073a1b59f0de5515580f635","started":true,"volumeMounts":[{"name":"kube-api-access-rnfmb","mountPath":"/var/run/secrets/kubernetes.io/serviceaccount","readOnly":true,"recursiveReadOnly":"Disabled"}]},{"name":"manager","state":{"running":{"startedAt":"2025-12-04T11:53:05Z"}},"lastState":{"terminated":{"exitCode":1,"reason":"Error","message":" 1 clustercatalog_controller.go:134] \"reconcile ending\" logger=\"catalogd-controller\" controller=\"clustercatalog\" controllerGroup=\"olm.operatorframework.io\" controllerKind=\"ClusterCatalog\" ClusterCatalog=\"openshift-redhat-marketplace\" namespace=\"\" name=\"openshift-redhat-marketplace\" reconcileID=\"fb1fdb7e-c9e9-4dcc-97f4-fadc44d75ef1\"\nI1204 11:48:04.735618 1 clustercatalog_controller.go:86] \"reconcile starting\" logger=\"catalogd-controller\" controller=\"clustercatalog\" controllerGroup=\"olm.operatorframework.io\" controllerKind=\"ClusterCatalog\" ClusterCatalog=\"openshift-redhat-operators\" namespace=\"\" name=\"openshift-redhat-operators\" reconcileID=\"00ae09fb-fa45-475d-8f10-ea42c7cb5b10\"\nI1204 11:48:04.735832 1 clustercatalog_controller.go:134] \"reconcile ending\" logger=\"catalogd-controller\" controller=\"clustercatalog\" controllerGroup=\"olm.operatorframework.io\" controllerKind=\"ClusterCatalog\" ClusterCatalog=\"openshift-redhat-operators\" namespace=\"\" name=\"openshift-redhat-operators\" reconcileID=\"00ae09fb-fa45-475d-8f10-ea42c7cb5b10\"\nI1204 11:48:07.688755 1 serverutil.go:96] \"handled request\" logger=\"catalogd-http-server\" host=\"10.128.0.41\" username=\"-\" method=\"GET\" uri=\"/catalogs/openshift-redhat-operators/api/v1/all\" protocol=\"HTTP/1.1\" status=200 size=13032258\nE1204 11:51:51.663609 1 leaderelection.go:429] Failed to update lock optimitically: Timeout: request did not complete within requested timeout - context deadline exceeded, falling back to slow path\nE1204 11:52:51.665290 1 leaderelection.go:436] error retrieving resource lock openshift-catalogd/catalogd-operator-lock: the server was unable to return a response in the time allotted, but may still be processing the request (get leases.coordination.k8s.io catalogd-operator-lock)\nI1204 11:53:04.660107 1 leaderelection.go:297] failed to renew lease openshift-catalogd/catalogd-operator-lock: timed out waiting for the condition\nE1204 11:53:04.660224 1 main.go:351] \"problem running manager\" err=\"leader election lost\" logger=\"setup\"\n","startedAt":"2025-12-04T11:41:38Z","finishedAt":"2025-12-04T11:53:04Z","containerID":"cri-o://51591cda8f462b13281f43dbc0b6afc384282884d691dc6e5a5e8196ffa850e4"}},"ready":true,"restartCount":2,"image":"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3f0aa9cd04713acc5c6fea721bd849e1500da8ae945e0b32000887f34d786e0b","imageID":"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3f0aa9cd04713acc5c6fea721bd849e1500da8ae945e0b32000887f34d786e0b","containerID":"cri-o://2d0a3b6f732b806f42b1873563371670f077c7370590e28e69e27a142e4a3367","started":true,"volumeMounts":[{"name":"cache","mountPath":"/var/cache/"},{"name":"catalogserver-certs","mountPath":"/var/certs"},{"name":"ca-certs","mountPath":"/var/ca-certs","readOnly":true,"recursiveReadOnly":"Disabled"},{"name":"etc-containers","mountPath":"/etc/containers","readOnly":true,"recursiveReadOnly":"Disabled"},{"name":"etc-docker","mountPath":"/etc/docker","readOnly":true,"recursiveReadOnly":"Disabled"},{"name":"kube-api-access-rnfmb","mountPath":"/var/run/secrets/kubernetes.io/serviceaccount","readOnly":true,"recursiveReadOnly":"Disabled"}]}],"qosClass":"Burstable"}}config/pod/openshift-apiserver-operator/openshift-apiserver-operator-7bf7f6b755-sh6qf.json0000640000000000000000000001640115114274061030270 0ustar0000000000000000{"metadata":{"name":"openshift-apiserver-operator-7bf7f6b755-sh6qf","generateName":"openshift-apiserver-operator-7bf7f6b755-","namespace":"openshift-apiserver-operator","uid":"a5b826c3-8638-4cf5-b974-b24e4282ce00","resourceVersion":"9863","creationTimestamp":"2025-12-04T11:35:48Z","labels":{"app":"openshift-apiserver-operator","pod-template-hash":"7bf7f6b755"},"annotations":{"k8s.ovn.org/pod-networks":"{\"default\":{\"ip_addresses\":[\"10.128.0.5/23\"],\"mac_address\":\"0a:58:0a:80:00:05\",\"gateway_ips\":[\"10.128.0.1\"],\"routes\":[{\"dest\":\"10.128.0.0/16\",\"nextHop\":\"10.128.0.1\"},{\"dest\":\"172.30.0.0/16\",\"nextHop\":\"10.128.0.1\"},{\"dest\":\"169.254.0.5/32\",\"nextHop\":\"10.128.0.1\"},{\"dest\":\"100.64.0.0/16\",\"nextHop\":\"10.128.0.1\"}],\"ip_address\":\"10.128.0.5/23\",\"gateway_ip\":\"10.128.0.1\",\"role\":\"primary\"}}","k8s.v1.cni.cncf.io/network-status":"[{\n \"name\": \"ovn-kubernetes\",\n \"interface\": \"eth0\",\n \"ips\": [\n \"10.128.0.5\"\n ],\n \"mac\": \"0a:58:0a:80:00:05\",\n \"default\": true,\n \"dns\": {}\n}]","openshift.io/required-scc":"nonroot-v2","openshift.io/scc":"nonroot-v2","seccomp.security.alpha.kubernetes.io/pod":"runtime/default"},"ownerReferences":[{"apiVersion":"apps/v1","kind":"ReplicaSet","name":"openshift-apiserver-operator-7bf7f6b755","uid":"7667a298-9ae4-401d-9367-9042d0ea9e56","controller":true,"blockOwnerDeletion":true}]},"spec":{"volumes":[{"name":"serving-cert","secret":{"secretName":"openshift-apiserver-operator-serving-cert","defaultMode":420,"optional":true}},{"name":"config","configMap":{"name":"openshift-apiserver-operator-config","defaultMode":420}},{"name":"kube-api-access-rhmlm","projected":{"sources":[{"serviceAccountToken":{"expirationSeconds":3607,"path":"token"}},{"configMap":{"name":"kube-root-ca.crt","items":[{"key":"ca.crt","path":"ca.crt"}]}},{"downwardAPI":{"items":[{"path":"namespace","fieldRef":{"apiVersion":"v1","fieldPath":"metadata.namespace"}}]}},{"configMap":{"name":"openshift-service-ca.crt","items":[{"key":"service-ca.crt","path":"service-ca.crt"}]}}],"defaultMode":420}}],"containers":[{"name":"openshift-apiserver-operator","image":"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8375671da86aa527ee7e291d86971b0baa823ffc7663b5a983084456e76c0f59","command":["cluster-openshift-apiserver-operator","operator"],"args":["--config=/var/run/configmaps/config/config.yaml"],"ports":[{"name":"metrics","containerPort":8443,"protocol":"TCP"}],"env":[{"name":"IMAGE","value":"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:df606f3b71d4376d1a2108c09f0d3dab455fc30bcb67c60e91590c105e9025bf"},{"name":"OPERATOR_IMAGE","value":"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8375671da86aa527ee7e291d86971b0baa823ffc7663b5a983084456e76c0f59"},{"name":"OPERATOR_IMAGE_VERSION","value":"4.18.29"},{"name":"OPERAND_IMAGE_VERSION","value":"4.18.29"},{"name":"KUBE_APISERVER_OPERATOR_IMAGE","value":"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:0f43c31aa3359159d4557dad3cfaf812d8ce44db9cb9ae970e06d3479070b660"}],"resources":{"requests":{"cpu":"10m","memory":"50Mi"}},"volumeMounts":[{"name":"config","mountPath":"/var/run/configmaps/config"},{"name":"serving-cert","mountPath":"/var/run/secrets/serving-cert"},{"name":"kube-api-access-rhmlm","readOnly":true,"mountPath":"/var/run/secrets/kubernetes.io/serviceaccount"}],"terminationMessagePath":"/dev/termination-log","terminationMessagePolicy":"FallbackToLogsOnError","imagePullPolicy":"IfNotPresent","securityContext":{"capabilities":{"drop":["ALL"]},"allowPrivilegeEscalation":false}}],"restartPolicy":"Always","terminationGracePeriodSeconds":30,"dnsPolicy":"ClusterFirst","nodeSelector":{"node-role.kubernetes.io/master":""},"serviceAccountName":"openshift-apiserver-operator","serviceAccount":"openshift-apiserver-operator","nodeName":"master-0","securityContext":{"seLinuxOptions":{"level":"s0:c14,c4"},"runAsUser":65534,"runAsNonRoot":true,"seccompProfile":{"type":"RuntimeDefault"}},"schedulerName":"default-scheduler","tolerations":[{"key":"node-role.kubernetes.io/master","operator":"Exists","effect":"NoSchedule"},{"key":"node-role.kubernetes.io/control-plane","operator":"Exists","effect":"NoExecute"},{"key":"node.kubernetes.io/unreachable","operator":"Exists","effect":"NoExecute","tolerationSeconds":120},{"key":"node.kubernetes.io/not-ready","operator":"Exists","effect":"NoExecute","tolerationSeconds":120},{"key":"node.kubernetes.io/memory-pressure","operator":"Exists","effect":"NoSchedule"}],"priorityClassName":"system-cluster-critical","priority":2000000000,"enableServiceLinks":true,"preemptionPolicy":"PreemptLowerPriority"},"status":{"phase":"Running","conditions":[{"type":"PodReadyToStartContainers","status":"True","lastProbeTime":null,"lastTransitionTime":"2025-12-04T11:37:34Z"},{"type":"Initialized","status":"True","lastProbeTime":null,"lastTransitionTime":"2025-12-04T11:37:30Z"},{"type":"Ready","status":"True","lastProbeTime":null,"lastTransitionTime":"2025-12-04T11:45:01Z"},{"type":"ContainersReady","status":"True","lastProbeTime":null,"lastTransitionTime":"2025-12-04T11:45:01Z"},{"type":"PodScheduled","status":"True","lastProbeTime":null,"lastTransitionTime":"2025-12-04T11:37:30Z"}],"hostIP":"192.168.32.10","hostIPs":[{"ip":"192.168.32.10"}],"podIP":"10.128.0.5","podIPs":[{"ip":"10.128.0.5"}],"startTime":"2025-12-04T11:37:30Z","containerStatuses":[{"name":"openshift-apiserver-operator","state":{"running":{"startedAt":"2025-12-04T11:43:20Z"}},"lastState":{"terminated":{"exitCode":255,"reason":"Error","message":"I1204 11:41:38.351624 1 cmd.go:253] Using service-serving-cert provided certificates\nI1204 11:41:38.351758 1 leaderelection.go:121] The leader election gives 4 retries and allows for 30s of clock skew. The kube-apiserver downtime tolerance is 78s. Worst non-graceful lease acquisition is 2m43s. Worst graceful lease acquisition is {26s}.\nI1204 11:41:38.352205 1 observer_polling.go:159] Starting file observer\nW1204 11:42:38.360993 1 builder.go:272] unable to get owner reference (falling back to namespace): the server was unable to return a response in the time allotted, but may still be processing the request (get pods)\nI1204 11:42:38.361519 1 builder.go:304] openshift-apiserver-operator version -\nF1204 11:43:08.595885 1 cmd.go:182] failed checking apiserver connectivity: Get \"https://172.30.0.1:443/apis/coordination.k8s.io/v1/namespaces/openshift-apiserver-operator/leases/openshift-apiserver-operator-lock\": context deadline exceeded\n","startedAt":"2025-12-04T11:41:38Z","finishedAt":"2025-12-04T11:43:08Z","containerID":"cri-o://bc1d5c88553d9975dfd695327c0f71abfbae1a4a9e554283b59af5847061176e"}},"ready":true,"restartCount":2,"image":"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8375671da86aa527ee7e291d86971b0baa823ffc7663b5a983084456e76c0f59","imageID":"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8375671da86aa527ee7e291d86971b0baa823ffc7663b5a983084456e76c0f59","containerID":"cri-o://29a193498b176ae14f0ca7dff8d35917ba4713a9868840310a0be48351f34521","started":true,"volumeMounts":[{"name":"config","mountPath":"/var/run/configmaps/config"},{"name":"serving-cert","mountPath":"/var/run/secrets/serving-cert"},{"name":"kube-api-access-rhmlm","mountPath":"/var/run/secrets/kubernetes.io/serviceaccount","readOnly":true,"recursiveReadOnly":"Disabled"}]}],"qosClass":"Burstable"}}config/configmaps/kube-system/cluster-config-v1/install-config0000640000000000000000000000155515114274061023216 0ustar0000000000000000apiVersion: v1 baseDomain: xxxxxxxxxxxxx compute: - hyperthreading: Enabled name: worker platform: {} replicas: 0 controlPlane: hyperthreading: Enabled name: master platform: {} replicas: 1 metadata: creationTimestamp: null name: sno networking: clusterNetwork: - cidr: 10.128.0.0/16 hostPrefix: 23 machineNetwork: - cidr: 192.168.32.0/24 networkType: OVNKubernetes serviceNetwork: - 172.30.0.0/16 platform: none: {} publish: External pullSecret: "" sshKey: xxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxx config/configmaps/openshift-config/openshift-service-ca.crt/service-ca.crt0000640000000000000000000000010115114274061025426 0ustar0000000000000000-----BEGIN CERTIFICATE----- ANONYMIZED -----END CERTIFICATE----- config/configmaps/openshift-config/openshift-install-manifests/version0000640000000000000000000000000715114274061025135 0ustar0000000000000000v4.18.0config/configmaps/openshift-config/openshift-install-manifests/invoker0000640000000000000000000000001715114274061025126 0ustar0000000000000000agent-installerconfig/configmaps/openshift-config/kube-root-ca.crt/ca.crt0000640000000000000000000000030315114274061022246 0ustar0000000000000000-----BEGIN CERTIFICATE----- ANONYMIZED -----END CERTIFICATE----- -----BEGIN CERTIFICATE----- ANONYMIZED -----END CERTIFICATE----- -----BEGIN CERTIFICATE----- ANONYMIZED -----END CERTIFICATE----- config/configmaps/openshift-config/initial-kube-apiserver-server-ca/ca-bundle.crt0000640000000000000000000000050515114274061026702 0ustar0000000000000000-----BEGIN CERTIFICATE----- ANONYMIZED -----END CERTIFICATE----- -----BEGIN CERTIFICATE----- ANONYMIZED -----END CERTIFICATE----- -----BEGIN CERTIFICATE----- ANONYMIZED -----END CERTIFICATE----- -----BEGIN CERTIFICATE----- ANONYMIZED -----END CERTIFICATE----- -----BEGIN CERTIFICATE----- ANONYMIZED -----END CERTIFICATE----- config/configmaps/openshift-config/etcd-serving-ca/ca-bundle.crt0000640000000000000000000000010115114274061023405 0ustar0000000000000000-----BEGIN CERTIFICATE----- ANONYMIZED -----END CERTIFICATE----- config/configmaps/openshift-config/etcd-ca-bundle/ca-bundle.crt0000640000000000000000000000010115114274061023201 0ustar0000000000000000-----BEGIN CERTIFICATE----- ANONYMIZED -----END CERTIFICATE----- config/configmaps/openshift-config/admin-kubeconfig-client-ca/ca-bundle.crt0000640000000000000000000000010115114274061025471 0ustar0000000000000000-----BEGIN CERTIFICATE----- ANONYMIZED -----END CERTIFICATE----- aggregated/unused_machine_configs_count.json0000640000000000000000000000004115114274061020130 0ustar0000000000000000{"unused_machineconfigs_count":9}events/openshift-cluster-version.json0000640000000000000000000002545215114274061016611 0ustar0000000000000000{"items":[{"namespace":"openshift-cluster-version","lastTimestamp":"2025-12-04T11:30:50Z","reason":"LeaderElection","message":"master-0_c61b9c51-689e-4779-8540-4c140f336d37 became leader","type":"Normal"},{"namespace":"openshift-cluster-version","lastTimestamp":"2025-12-04T11:30:52Z","reason":"ScalingReplicaSet","message":"Scaled up replica set cluster-version-operator-77dfcc565f to 1","type":"Normal"},{"namespace":"openshift-cluster-version","lastTimestamp":"2025-12-04T11:30:53Z","reason":"LeaderElection","message":"master-0_87cbe5e2-b68d-44c6-9a7c-9634cfb7073d became leader","type":"Normal"},{"namespace":"openshift-cluster-version","lastTimestamp":"2025-12-04T11:31:10Z","reason":"LeaderElection","message":"master-0_90606d2c-c56e-4c2c-bba2-d1d365b5695f became leader","type":"Normal"},{"namespace":"openshift-cluster-version","lastTimestamp":"2025-12-04T11:31:12Z","reason":"RetrievePayload","message":"Retrieving and verifying payload version=\"4.18.29\" image=\"quay.io/openshift-release-dev/ocp-release@sha256:8c885ea0b3c5124989f0a9b93eba98eb9fca6bbd0262772d85d90bf713a4d572\"","type":"Normal"},{"namespace":"openshift-cluster-version","lastTimestamp":"2025-12-04T11:31:12Z","reason":"LoadPayload","message":"Loading payload version=\"4.18.29\" image=\"quay.io/openshift-release-dev/ocp-release@sha256:8c885ea0b3c5124989f0a9b93eba98eb9fca6bbd0262772d85d90bf713a4d572\"","type":"Normal"},{"namespace":"openshift-cluster-version","lastTimestamp":"2025-12-04T11:31:13Z","reason":"PayloadLoaded","message":"Payload loaded version=\"4.18.29\" image=\"quay.io/openshift-release-dev/ocp-release@sha256:8c885ea0b3c5124989f0a9b93eba98eb9fca6bbd0262772d85d90bf713a4d572\" architecture=\"amd64\"","type":"Normal"},{"namespace":"openshift-cluster-version","lastTimestamp":"2025-12-04T11:31:33Z","reason":"FailedCreate","message":"Error creating: pods \"cluster-version-operator-77dfcc565f-\" is forbidden: autoscaling.openshift.io/ManagementCPUsOverride the cluster does not have any nodes","type":"Warning"},{"namespace":"openshift-cluster-version","lastTimestamp":"2025-12-04T11:35:47Z","reason":"FailedCreate","message":"Error creating: pods \"cluster-version-operator-77dfcc565f-\" is forbidden: autoscaling.openshift.io/ManagementCPUsOverride the cluster does not have any nodes","type":"Warning"},{"namespace":"openshift-cluster-version","lastTimestamp":"2025-12-04T11:35:49Z","reason":"Scheduled","message":"Successfully assigned openshift-cluster-version/cluster-version-operator-77dfcc565f-nqpsd to master-0","type":"Normal"},{"namespace":"openshift-cluster-version","lastTimestamp":"2025-12-04T11:35:49Z","reason":"SuccessfulCreate","message":"Created pod: cluster-version-operator-77dfcc565f-nqpsd","type":"Normal"},{"namespace":"openshift-cluster-version","lastTimestamp":"2025-12-04T11:37:18Z","reason":"FailedMount","message":"MountVolume.SetUp failed for volume \"serving-cert\" : secret \"cluster-version-operator-serving-cert\" not found","type":"Warning"},{"namespace":"openshift-cluster-version","lastTimestamp":"2025-12-04T11:37:50Z","reason":"FailedMount","message":"MountVolume.SetUp failed for volume \"serving-cert\" : secret \"cluster-version-operator-serving-cert\" not found","type":"Warning"},{"namespace":"openshift-cluster-version","lastTimestamp":"2025-12-04T11:37:54Z","reason":"Pulling","message":"Pulling image \"quay.io/openshift-release-dev/ocp-release@sha256:8c885ea0b3c5124989f0a9b93eba98eb9fca6bbd0262772d85d90bf713a4d572\"","type":"Normal"},{"namespace":"openshift-cluster-version","lastTimestamp":"2025-12-04T11:37:57Z","reason":"Started","message":"Started container cluster-version-operator","type":"Normal"},{"namespace":"openshift-cluster-version","lastTimestamp":"2025-12-04T11:37:57Z","reason":"Created","message":"Created container: cluster-version-operator","type":"Normal"},{"namespace":"openshift-cluster-version","lastTimestamp":"2025-12-04T11:37:57Z","reason":"Pulled","message":"Successfully pulled image \"quay.io/openshift-release-dev/ocp-release@sha256:8c885ea0b3c5124989f0a9b93eba98eb9fca6bbd0262772d85d90bf713a4d572\" in 3.053s (3.053s including waiting). Image size: 512452153 bytes.","type":"Normal"},{"namespace":"openshift-cluster-version","lastTimestamp":"2025-12-04T11:37:58Z","reason":"LeaderElection","message":"master-0_369556f3-f56b-4ce4-81c9-ae023ba4b139 became leader","type":"Normal"},{"namespace":"openshift-cluster-version","lastTimestamp":"2025-12-04T11:37:59Z","reason":"LoadPayload","message":"Loading payload version=\"4.18.29\" image=\"quay.io/openshift-release-dev/ocp-release@sha256:8c885ea0b3c5124989f0a9b93eba98eb9fca6bbd0262772d85d90bf713a4d572\"","type":"Normal"},{"namespace":"openshift-cluster-version","lastTimestamp":"2025-12-04T11:37:59Z","reason":"RetrievePayload","message":"Retrieving and verifying payload version=\"4.18.29\" image=\"quay.io/openshift-release-dev/ocp-release@sha256:8c885ea0b3c5124989f0a9b93eba98eb9fca6bbd0262772d85d90bf713a4d572\"","type":"Normal"},{"namespace":"openshift-cluster-version","lastTimestamp":"2025-12-04T11:38:01Z","reason":"PayloadLoaded","message":"Payload loaded version=\"4.18.29\" image=\"quay.io/openshift-release-dev/ocp-release@sha256:8c885ea0b3c5124989f0a9b93eba98eb9fca6bbd0262772d85d90bf713a4d572\" architecture=\"amd64\"","type":"Normal"},{"namespace":"openshift-cluster-version","lastTimestamp":"2025-12-04T11:38:18Z","reason":"SuccessfulDelete","message":"Deleted pod: cluster-version-operator-77dfcc565f-nqpsd","type":"Normal"},{"namespace":"openshift-cluster-version","lastTimestamp":"2025-12-04T11:38:18Z","reason":"LeaderElection","message":"master-0_369556f3-f56b-4ce4-81c9-ae023ba4b139 stopped leading","type":"Normal"},{"namespace":"openshift-cluster-version","lastTimestamp":"2025-12-04T11:38:18Z","reason":"ScalingReplicaSet","message":"Scaled down replica set cluster-version-operator-77dfcc565f to 0 from 1","type":"Normal"},{"namespace":"openshift-cluster-version","lastTimestamp":"2025-12-04T11:38:18Z","reason":"Killing","message":"Stopping container cluster-version-operator","type":"Normal"},{"namespace":"openshift-cluster-version","lastTimestamp":"2025-12-04T11:38:20Z","reason":"SuccessfulCreate","message":"Created pod: cluster-version-operator-6d5d5dcc89-cw2hx","type":"Normal"},{"namespace":"openshift-cluster-version","lastTimestamp":"2025-12-04T11:38:20Z","reason":"ScalingReplicaSet","message":"Scaled up replica set cluster-version-operator-6d5d5dcc89 to 1","type":"Normal"},{"namespace":"openshift-cluster-version","lastTimestamp":"2025-12-04T11:38:20Z","reason":"Scheduled","message":"Successfully assigned openshift-cluster-version/cluster-version-operator-6d5d5dcc89-cw2hx to master-0","type":"Normal"},{"namespace":"openshift-cluster-version","lastTimestamp":"2025-12-04T11:38:21Z","reason":"LeaderElection","message":"master-0_24c3cd95-5f05-4668-945c-5fee4fae08e7 became leader","type":"Normal"},{"namespace":"openshift-cluster-version","lastTimestamp":"2025-12-04T11:38:23Z","reason":"LoadPayload","message":"Loading payload version=\"4.18.29\" image=\"quay.io/openshift-release-dev/ocp-release@sha256:8c885ea0b3c5124989f0a9b93eba98eb9fca6bbd0262772d85d90bf713a4d572\"","type":"Normal"},{"namespace":"openshift-cluster-version","lastTimestamp":"2025-12-04T11:38:23Z","reason":"RetrievePayload","message":"Retrieving and verifying payload version=\"4.18.29\" image=\"quay.io/openshift-release-dev/ocp-release@sha256:8c885ea0b3c5124989f0a9b93eba98eb9fca6bbd0262772d85d90bf713a4d572\"","type":"Normal"},{"namespace":"openshift-cluster-version","lastTimestamp":"2025-12-04T11:38:25Z","reason":"PayloadLoaded","message":"Payload loaded version=\"4.18.29\" image=\"quay.io/openshift-release-dev/ocp-release@sha256:8c885ea0b3c5124989f0a9b93eba98eb9fca6bbd0262772d85d90bf713a4d572\" architecture=\"amd64\"","type":"Normal"},{"namespace":"openshift-cluster-version","lastTimestamp":"2025-12-04T11:43:55Z","reason":"LeaderElection","message":"master-0_24c3cd95-5f05-4668-945c-5fee4fae08e7 stopped leading","type":"Normal"},{"namespace":"openshift-cluster-version","lastTimestamp":"2025-12-04T11:50:18Z","reason":"LeaderElection","message":"master-0_5b5f4e61-8577-4c39-b810-b785e09ab344 became leader","type":"Normal"},{"namespace":"openshift-cluster-version","lastTimestamp":"2025-12-04T11:50:20Z","reason":"LoadPayload","message":"Loading payload version=\"4.18.29\" image=\"quay.io/openshift-release-dev/ocp-release@sha256:8c885ea0b3c5124989f0a9b93eba98eb9fca6bbd0262772d85d90bf713a4d572\"","type":"Normal"},{"namespace":"openshift-cluster-version","lastTimestamp":"2025-12-04T11:50:20Z","reason":"RetrievePayload","message":"Retrieving and verifying payload version=\"4.18.29\" image=\"quay.io/openshift-release-dev/ocp-release@sha256:8c885ea0b3c5124989f0a9b93eba98eb9fca6bbd0262772d85d90bf713a4d572\"","type":"Normal"},{"namespace":"openshift-cluster-version","lastTimestamp":"2025-12-04T11:50:21Z","reason":"PayloadLoaded","message":"Payload loaded version=\"4.18.29\" image=\"quay.io/openshift-release-dev/ocp-release@sha256:8c885ea0b3c5124989f0a9b93eba98eb9fca6bbd0262772d85d90bf713a4d572\" architecture=\"amd64\"","type":"Normal"},{"namespace":"openshift-cluster-version","lastTimestamp":"2025-12-04T11:55:46Z","reason":"LeaderElection","message":"master-0_5b5f4e61-8577-4c39-b810-b785e09ab344 stopped leading","type":"Normal"},{"namespace":"openshift-cluster-version","lastTimestamp":"2025-12-04T11:56:26Z","reason":"Pulled","message":"Container image \"quay.io/openshift-release-dev/ocp-release@sha256:8c885ea0b3c5124989f0a9b93eba98eb9fca6bbd0262772d85d90bf713a4d572\" already present on machine","type":"Normal"},{"namespace":"openshift-cluster-version","lastTimestamp":"2025-12-04T11:56:26Z","reason":"Created","message":"Created container: cluster-version-operator","type":"Normal"},{"namespace":"openshift-cluster-version","lastTimestamp":"2025-12-04T11:56:26Z","reason":"Started","message":"Started container cluster-version-operator","type":"Normal"},{"namespace":"openshift-cluster-version","lastTimestamp":"2025-12-04T11:56:26Z","reason":"LeaderElection","message":"master-0_c465a515-4df4-43c3-8d97-3c391bd274ae became leader","type":"Normal"},{"namespace":"openshift-cluster-version","lastTimestamp":"2025-12-04T11:56:28Z","reason":"RetrievePayload","message":"Retrieving and verifying payload version=\"4.18.29\" image=\"quay.io/openshift-release-dev/ocp-release@sha256:8c885ea0b3c5124989f0a9b93eba98eb9fca6bbd0262772d85d90bf713a4d572\"","type":"Normal"},{"namespace":"openshift-cluster-version","lastTimestamp":"2025-12-04T11:56:28Z","reason":"LoadPayload","message":"Loading payload version=\"4.18.29\" image=\"quay.io/openshift-release-dev/ocp-release@sha256:8c885ea0b3c5124989f0a9b93eba98eb9fca6bbd0262772d85d90bf713a4d572\"","type":"Normal"},{"namespace":"openshift-cluster-version","lastTimestamp":"2025-12-04T11:56:29Z","reason":"PayloadLoaded","message":"Payload loaded version=\"4.18.29\" image=\"quay.io/openshift-release-dev/ocp-release@sha256:8c885ea0b3c5124989f0a9b93eba98eb9fca6bbd0262772d85d90bf713a4d572\" architecture=\"amd64\"","type":"Normal"}]}config/pod/openshift-cluster-version/cluster-version-operator-6d5d5dcc89-cw2hx.json0000640000000000000000000001374115114274061027040 0ustar0000000000000000{"metadata":{"name":"cluster-version-operator-6d5d5dcc89-cw2hx","generateName":"cluster-version-operator-6d5d5dcc89-","namespace":"openshift-cluster-version","uid":"a399a87b-8153-4045-a8d3-63fd88ae280f","resourceVersion":"11330","creationTimestamp":"2025-12-04T11:38:20Z","labels":{"k8s-app":"cluster-version-operator","pod-template-hash":"6d5d5dcc89"},"annotations":{"openshift.io/required-scc":"hostaccess","openshift.io/scc":"hostaccess"},"ownerReferences":[{"apiVersion":"apps/v1","kind":"ReplicaSet","name":"cluster-version-operator-6d5d5dcc89","uid":"3a6df863-f92d-455e-8869-68e0051920d9","controller":true,"blockOwnerDeletion":true}]},"spec":{"volumes":[{"name":"etc-ssl-certs","hostPath":{"path":"/etc/ssl/certs","type":""}},{"name":"etc-cvo-updatepayloads","hostPath":{"path":"/etc/cvo/updatepayloads","type":""}},{"name":"serving-cert","secret":{"secretName":"cluster-version-operator-serving-cert","defaultMode":420}},{"name":"service-ca","configMap":{"name":"openshift-service-ca.crt","defaultMode":420}},{"name":"kube-api-access","projected":{"sources":[{"serviceAccountToken":{"expirationSeconds":3600,"path":"token"}},{"configMap":{"name":"kube-root-ca.crt","items":[{"key":"ca.crt","path":"ca.crt"}]}},{"downwardAPI":{"items":[{"path":"namespace","fieldRef":{"apiVersion":"v1","fieldPath":"metadata.namespace"}}]}}],"defaultMode":420}}],"containers":[{"name":"cluster-version-operator","image":"quay.io/openshift-release-dev/ocp-release@sha256:8c885ea0b3c5124989f0a9b93eba98eb9fca6bbd0262772d85d90bf713a4d572","args":["start","--release-image=quay.io/openshift-release-dev/ocp-release@sha256:8c885ea0b3c5124989f0a9b93eba98eb9fca6bbd0262772d85d90bf713a4d572","--enable-auto-update=false","--listen=0.0.0.0:9099","--serving-cert-file=/etc/tls/serving-cert/tls.crt","--serving-key-file=/etc/tls/serving-cert/tls.key","--v=2","--always-enable-capabilities=Ingress"],"env":[{"name":"OPERATOR_IMAGE_VERSION","value":"0.0.1-snapshot"},{"name":"KUBERNETES_SERVICE_PORT","value":"6443"},{"name":"KUBERNETES_SERVICE_HOST","value":"api-int.sno.openstack.lab"},{"name":"NODE_NAME","valueFrom":{"fieldRef":{"apiVersion":"v1","fieldPath":"spec.nodeName"}}},{"name":"CLUSTER_PROFILE","value":"self-managed-high-availability"}],"resources":{"requests":{"cpu":"20m","memory":"50Mi"}},"volumeMounts":[{"name":"etc-ssl-certs","readOnly":true,"mountPath":"/etc/ssl/certs"},{"name":"etc-cvo-updatepayloads","readOnly":true,"mountPath":"/etc/cvo/updatepayloads"},{"name":"serving-cert","readOnly":true,"mountPath":"/etc/tls/serving-cert"},{"name":"service-ca","readOnly":true,"mountPath":"/etc/tls/service-ca"},{"name":"kube-api-access","readOnly":true,"mountPath":"/var/run/secrets/kubernetes.io/serviceaccount"}],"terminationMessagePath":"/dev/termination-log","terminationMessagePolicy":"FallbackToLogsOnError","imagePullPolicy":"IfNotPresent","securityContext":{"capabilities":{"drop":["KILL","MKNOD","SETGID","SETUID"]},"runAsUser":1000040000,"runAsNonRoot":true}}],"restartPolicy":"Always","terminationGracePeriodSeconds":130,"dnsPolicy":"Default","nodeSelector":{"node-role.kubernetes.io/master":""},"serviceAccountName":"default","serviceAccount":"default","automountServiceAccountToken":false,"nodeName":"master-0","hostNetwork":true,"securityContext":{"seLinuxOptions":{"level":"s0:c6,c5"},"fsGroup":1000040000},"schedulerName":"default-scheduler","tolerations":[{"key":"node-role.kubernetes.io/master","operator":"Exists","effect":"NoSchedule"},{"key":"node.kubernetes.io/network-unavailable","operator":"Exists","effect":"NoSchedule"},{"key":"node.kubernetes.io/not-ready","operator":"Exists","effect":"NoSchedule"},{"key":"node.kubernetes.io/unreachable","operator":"Exists","effect":"NoExecute","tolerationSeconds":120},{"key":"node.kubernetes.io/not-ready","operator":"Exists","effect":"NoExecute","tolerationSeconds":120},{"key":"node.kubernetes.io/memory-pressure","operator":"Exists","effect":"NoSchedule"}],"priorityClassName":"system-cluster-critical","priority":2000000000,"enableServiceLinks":true,"preemptionPolicy":"PreemptLowerPriority"},"status":{"phase":"Running","conditions":[{"type":"PodReadyToStartContainers","status":"True","lastProbeTime":null,"lastTransitionTime":"2025-12-04T11:38:22Z"},{"type":"Initialized","status":"True","lastProbeTime":null,"lastTransitionTime":"2025-12-04T11:38:20Z"},{"type":"Ready","status":"True","lastProbeTime":null,"lastTransitionTime":"2025-12-04T11:56:27Z"},{"type":"ContainersReady","status":"True","lastProbeTime":null,"lastTransitionTime":"2025-12-04T11:56:27Z"},{"type":"PodScheduled","status":"True","lastProbeTime":null,"lastTransitionTime":"2025-12-04T11:38:20Z"}],"hostIP":"192.168.32.10","hostIPs":[{"ip":"192.168.32.10"}],"podIP":"192.168.32.10","podIPs":[{"ip":"192.168.32.10"}],"startTime":"2025-12-04T11:38:20Z","containerStatuses":[{"name":"cluster-version-operator","state":{"running":{"startedAt":"2025-12-04T11:56:26Z"}},"lastState":{"terminated":{"exitCode":0,"reason":"Completed","startedAt":"2025-12-04T11:45:02Z","finishedAt":"2025-12-04T11:56:26Z","containerID":"cri-o://8203c30c9d158150398a01d336c4969a20aab618396e7ab3bc78d00c326809cc"}},"ready":true,"restartCount":2,"image":"quay.io/openshift-release-dev/ocp-release@sha256:8c885ea0b3c5124989f0a9b93eba98eb9fca6bbd0262772d85d90bf713a4d572","imageID":"quay.io/openshift-release-dev/ocp-release@sha256:8c885ea0b3c5124989f0a9b93eba98eb9fca6bbd0262772d85d90bf713a4d572","containerID":"cri-o://ca65c28331bad9eec792ccc54a97b6670a456f91f8ece05a18970c8e1f00c399","started":true,"volumeMounts":[{"name":"etc-ssl-certs","mountPath":"/etc/ssl/certs","readOnly":true,"recursiveReadOnly":"Disabled"},{"name":"etc-cvo-updatepayloads","mountPath":"/etc/cvo/updatepayloads","readOnly":true,"recursiveReadOnly":"Disabled"},{"name":"serving-cert","mountPath":"/etc/tls/serving-cert","readOnly":true,"recursiveReadOnly":"Disabled"},{"name":"service-ca","mountPath":"/etc/tls/service-ca","readOnly":true,"recursiveReadOnly":"Disabled"},{"name":"kube-api-access","mountPath":"/var/run/secrets/kubernetes.io/serviceaccount","readOnly":true,"recursiveReadOnly":"Disabled"}]}],"qosClass":"Burstable"}}config/id0000640000000000000000000000004415114274061011023 0ustar0000000000000000741e5bbb-28e6-452d-a27b-b2222f825f7aconfig/version.json0000640000000000000000000000574315114274061013077 0ustar0000000000000000{"metadata":{"name":"version","uid":"a151bf9a-21b6-4d33-89ec-0e4f83e8eab6","resourceVersion":"14226","generation":2,"creationTimestamp":"2025-12-04T11:30:39Z"},"spec":{"clusterID":"741e5bbb-28e6-452d-a27b-b2222f825f7a","channel":"stable-4.18","signatureStores":null},"status":{"desired":{"version":"4.18.29","image":"quay.io/openshift-release-dev/ocp-release@sha256:8c885ea0b3c5124989f0a9b93eba98eb9fca6bbd0262772d85d90bf713a4d572","url":"https://access.redhat.com/errata/RHBA-2025:21797","channels":["candidate-4.18","candidate-4.19","candidate-4.20","eus-4.18","fast-4.18","fast-4.19","fast-4.20","stable-4.18"]},"history":[{"state":"Partial","startedTime":"2025-12-04T11:31:13Z","completionTime":null,"version":"4.18.29","image":"quay.io/openshift-release-dev/ocp-release@sha256:8c885ea0b3c5124989f0a9b93eba98eb9fca6bbd0262772d85d90bf713a4d572","verified":false}],"observedGeneration":2,"versionHash":"PLBD6Tt9bD0=","capabilities":{"enabledCapabilities":["Build","CSISnapshot","CloudControllerManager","CloudCredential","Console","DeploymentConfig","ImageRegistry","Ingress","Insights","MachineAPI","NodeTuning","OperatorLifecycleManager","OperatorLifecycleManagerV1","Storage","baremetal","marketplace","openshift-samples"],"knownCapabilities":["Build","CSISnapshot","CloudControllerManager","CloudCredential","Console","DeploymentConfig","ImageRegistry","Ingress","Insights","MachineAPI","NodeTuning","OperatorLifecycleManager","OperatorLifecycleManagerV1","Storage","baremetal","marketplace","openshift-samples"]},"conditions":[{"type":"RetrievedUpdates","status":"True","lastTransitionTime":"2025-12-04T11:31:13Z"},{"type":"Upgradeable","status":"False","lastTransitionTime":"2025-12-04T11:31:13Z","reason":"AdminAckRequired","message":"Kubernetes 1.32 and therefore OpenShift 4.19 remove several APIs which require admin consideration. Please see the knowledge article https://access.redhat.com/articles/7112216 for details and instructions."},{"type":"ImplicitlyEnabledCapabilities","status":"False","lastTransitionTime":"2025-12-04T11:31:13Z","reason":"AsExpected","message":"Capabilities match configured spec"},{"type":"ReleaseAccepted","status":"True","lastTransitionTime":"2025-12-04T11:31:13Z","reason":"PayloadLoaded","message":"Payload loaded version=\"4.18.29\" image=\"quay.io/openshift-release-dev/ocp-release@sha256:8c885ea0b3c5124989f0a9b93eba98eb9fca6bbd0262772d85d90bf713a4d572\" architecture=\"amd64\""},{"type":"Available","status":"False","lastTransitionTime":"2025-12-04T11:31:13Z"},{"type":"Failing","status":"True","lastTransitionTime":"2025-12-04T12:01:14Z","reason":"ClusterOperatorsNotAvailable","message":"Cluster operators authentication, console, dns, kube-controller-manager, monitoring, openshift-controller-manager, operator-lifecycle-manager-packageserver are not available"},{"type":"Progressing","status":"True","lastTransitionTime":"2025-12-04T11:31:13Z","reason":"ClusterOperatorsNotAvailable","message":"Unable to apply 4.18.29: some cluster operators are not available"}],"availableUpdates":null}}aggregated/ingress_controllers_certs.json0000640000000000000000000000045715114274061017534 0ustar0000000000000000[{"name":"router-ca","namespace":"openshift-ingress-operator","not_before":"2025-12-04T11:38:19Z","not_after":"2027-12-04T11:38:20Z","controllers":[]},{"name":"router-certs-default","namespace":"openshift-ingress","not_before":"2025-12-04T11:38:21Z","not_after":"2027-12-04T11:38:22Z","controllers":[]}]config/crd/volumesnapshotcontents.snapshot.storage.k8s.io.json0000640000000000000000000005444315114274061023443 0ustar0000000000000000{"metadata":{"name":"volumesnapshotcontents.snapshot.storage.k8s.io","uid":"17839266-4ac1-45f0-b562-477d6fd74f38","resourceVersion":"3988","generation":1,"creationTimestamp":"2025-12-04T11:37:44Z","annotations":{"api-approved.kubernetes.io":"https://github.com/kubernetes-csi/external-snapshotter/pull/955","controller-gen.kubebuilder.io/version":"v0.15.0"}},"spec":{"group":"snapshot.storage.k8s.io","names":{"plural":"volumesnapshotcontents","singular":"volumesnapshotcontent","shortNames":["vsc","vscs"],"kind":"VolumeSnapshotContent","listKind":"VolumeSnapshotContentList"},"scope":"Cluster","versions":[{"name":"v1","served":true,"storage":true,"schema":{"openAPIV3Schema":{"description":"VolumeSnapshotContent represents the actual \"on-disk\" snapshot object in the\nunderlying storage system","type":"object","required":["spec"],"properties":{"apiVersion":{"description":"APIVersion defines the versioned schema of this representation of an object.\nServers should convert recognized schemas to the latest internal value, and\nmay reject unrecognized values.\nMore info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#resources","type":"string"},"kind":{"description":"Kind is a string value representing the REST resource this object represents.\nServers may infer this from the endpoint the client submits requests to.\nCannot be updated.\nIn CamelCase.\nMore info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#types-kinds","type":"string"},"metadata":{"type":"object"},"spec":{"description":"spec defines properties of a VolumeSnapshotContent created by the underlying storage system.\nRequired.","type":"object","required":["deletionPolicy","driver","source","volumeSnapshotRef"],"properties":{"deletionPolicy":{"description":"deletionPolicy determines whether this VolumeSnapshotContent and its physical snapshot on\nthe underlying storage system should be deleted when its bound VolumeSnapshot is deleted.\nSupported values are \"Retain\" and \"Delete\".\n\"Retain\" means that the VolumeSnapshotContent and its physical snapshot on underlying storage system are kept.\n\"Delete\" means that the VolumeSnapshotContent and its physical snapshot on underlying storage system are deleted.\nFor dynamically provisioned snapshots, this field will automatically be filled in by the\nCSI snapshotter sidecar with the \"DeletionPolicy\" field defined in the corresponding\nVolumeSnapshotClass.\nFor pre-existing snapshots, users MUST specify this field when creating the\n VolumeSnapshotContent object.\nRequired.","type":"string","enum":["Delete","Retain"]},"driver":{"description":"driver is the name of the CSI driver used to create the physical snapshot on\nthe underlying storage system.\nThis MUST be the same as the name returned by the CSI GetPluginName() call for\nthat driver.\nRequired.","type":"string"},"source":{"description":"source specifies whether the snapshot is (or should be) dynamically provisioned\nor already exists, and just requires a Kubernetes object representation.\nThis field is immutable after creation.\nRequired.","type":"object","properties":{"snapshotHandle":{"description":"snapshotHandle specifies the CSI \"snapshot_id\" of a pre-existing snapshot on\nthe underlying storage system for which a Kubernetes object representation\nwas (or should be) created.\nThis field is immutable.","type":"string","x-kubernetes-validations":[{"rule":"self == oldSelf","message":"snapshotHandle is immutable"}]},"volumeHandle":{"description":"volumeHandle specifies the CSI \"volume_id\" of the volume from which a snapshot\nshould be dynamically taken from.\nThis field is immutable.","type":"string","x-kubernetes-validations":[{"rule":"self == oldSelf","message":"volumeHandle is immutable"}]}},"x-kubernetes-validations":[{"rule":"!has(oldSelf.volumeHandle) || has(self.volumeHandle)","message":"volumeHandle is required once set"},{"rule":"!has(oldSelf.snapshotHandle) || has(self.snapshotHandle)","message":"snapshotHandle is required once set"},{"rule":"(has(self.volumeHandle) \u0026\u0026 !has(self.snapshotHandle)) || (!has(self.volumeHandle) \u0026\u0026 has(self.snapshotHandle))","message":"exactly one of volumeHandle and snapshotHandle must be set"}]},"sourceVolumeMode":{"description":"SourceVolumeMode is the mode of the volume whose snapshot is taken.\nCan be either “Filesystem” or “Block”.\nIf not specified, it indicates the source volume's mode is unknown.\nThis field is immutable.\nThis field is an alpha field.","type":"string","x-kubernetes-validations":[{"rule":"self == oldSelf","message":"sourceVolumeMode is immutable"}]},"volumeSnapshotClassName":{"description":"name of the VolumeSnapshotClass from which this snapshot was (or will be)\ncreated.\nNote that after provisioning, the VolumeSnapshotClass may be deleted or\nrecreated with different set of values, and as such, should not be referenced\npost-snapshot creation.","type":"string"},"volumeSnapshotRef":{"description":"volumeSnapshotRef specifies the VolumeSnapshot object to which this\nVolumeSnapshotContent object is bound.\nVolumeSnapshot.Spec.VolumeSnapshotContentName field must reference to\nthis VolumeSnapshotContent's name for the bidirectional binding to be valid.\nFor a pre-existing VolumeSnapshotContent object, name and namespace of the\nVolumeSnapshot object MUST be provided for binding to happen.\nThis field is immutable after creation.\nRequired.","type":"object","properties":{"apiVersion":{"description":"API version of the referent.","type":"string"},"fieldPath":{"description":"If referring to a piece of an object instead of an entire object, this string\nshould contain a valid JSON/Go field access statement, such as desiredState.manifest.containers[2].\nFor example, if the object reference is to a container within a pod, this would take on a value like:\n\"spec.containers{name}\" (where \"name\" refers to the name of the container that triggered\nthe event) or if no container name is specified \"spec.containers[2]\" (container with\nindex 2 in this pod). This syntax is chosen only to have some well-defined way of\nreferencing a part of an object.\nTODO: this design is not final and this field is subject to change in the future.","type":"string"},"kind":{"description":"Kind of the referent.\nMore info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#types-kinds","type":"string"},"name":{"description":"Name of the referent.\nMore info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names","type":"string"},"namespace":{"description":"Namespace of the referent.\nMore info: https://kubernetes.io/docs/concepts/overview/working-with-objects/namespaces/","type":"string"},"resourceVersion":{"description":"Specific resourceVersion to which this reference is made, if any.\nMore info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#concurrency-control-and-consistency","type":"string"},"uid":{"description":"UID of the referent.\nMore info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#uids","type":"string"}},"x-kubernetes-map-type":"atomic","x-kubernetes-validations":[{"rule":"has(self.name) \u0026\u0026 has(self.__namespace__)","message":"both spec.volumeSnapshotRef.name and spec.volumeSnapshotRef.namespace must be set"}]}},"x-kubernetes-validations":[{"rule":"!has(oldSelf.sourceVolumeMode) || has(self.sourceVolumeMode)","message":"sourceVolumeMode is required once set"}]},"status":{"description":"status represents the current information of a snapshot.","type":"object","properties":{"creationTime":{"description":"creationTime is the timestamp when the point-in-time snapshot is taken\nby the underlying storage system.\nIn dynamic snapshot creation case, this field will be filled in by the\nCSI snapshotter sidecar with the \"creation_time\" value returned from CSI\n\"CreateSnapshot\" gRPC call.\nFor a pre-existing snapshot, this field will be filled with the \"creation_time\"\nvalue returned from the CSI \"ListSnapshots\" gRPC call if the driver supports it.\nIf not specified, it indicates the creation time is unknown.\nThe format of this field is a Unix nanoseconds time encoded as an int64.\nOn Unix, the command `date +%s%N` returns the current time in nanoseconds\nsince 1970-01-01 00:00:00 UTC.","type":"integer","format":"int64"},"error":{"description":"error is the last observed error during snapshot creation, if any.\nUpon success after retry, this error field will be cleared.","type":"object","properties":{"message":{"description":"message is a string detailing the encountered error during snapshot\ncreation if specified.\nNOTE: message may be logged, and it should not contain sensitive\ninformation.","type":"string"},"time":{"description":"time is the timestamp when the error was encountered.","type":"string","format":"date-time"}}},"readyToUse":{"description":"readyToUse indicates if a snapshot is ready to be used to restore a volume.\nIn dynamic snapshot creation case, this field will be filled in by the\nCSI snapshotter sidecar with the \"ready_to_use\" value returned from CSI\n\"CreateSnapshot\" gRPC call.\nFor a pre-existing snapshot, this field will be filled with the \"ready_to_use\"\nvalue returned from the CSI \"ListSnapshots\" gRPC call if the driver supports it,\notherwise, this field will be set to \"True\".\nIf not specified, it means the readiness of a snapshot is unknown.","type":"boolean"},"restoreSize":{"description":"restoreSize represents the complete size of the snapshot in bytes.\nIn dynamic snapshot creation case, this field will be filled in by the\nCSI snapshotter sidecar with the \"size_bytes\" value returned from CSI\n\"CreateSnapshot\" gRPC call.\nFor a pre-existing snapshot, this field will be filled with the \"size_bytes\"\nvalue returned from the CSI \"ListSnapshots\" gRPC call if the driver supports it.\nWhen restoring a volume from this snapshot, the size of the volume MUST NOT\nbe smaller than the restoreSize if it is specified, otherwise the restoration will fail.\nIf not specified, it indicates that the size is unknown.","type":"integer","format":"int64","minimum":0},"snapshotHandle":{"description":"snapshotHandle is the CSI \"snapshot_id\" of a snapshot on the underlying storage system.\nIf not specified, it indicates that dynamic snapshot creation has either failed\nor it is still in progress.","type":"string"},"volumeGroupSnapshotHandle":{"description":"VolumeGroupSnapshotHandle is the CSI \"group_snapshot_id\" of a group snapshot\non the underlying storage system.","type":"string"}}}}}},"subresources":{"status":{}},"additionalPrinterColumns":[{"name":"ReadyToUse","type":"boolean","description":"Indicates if the snapshot is ready to be used to restore a volume.","jsonPath":".status.readyToUse"},{"name":"RestoreSize","type":"integer","description":"Represents the complete size of the snapshot in bytes","jsonPath":".status.restoreSize"},{"name":"DeletionPolicy","type":"string","description":"Determines whether this VolumeSnapshotContent and its physical snapshot on the underlying storage system should be deleted when its bound VolumeSnapshot is deleted.","jsonPath":".spec.deletionPolicy"},{"name":"Driver","type":"string","description":"Name of the CSI driver used to create the physical snapshot on the underlying storage system.","jsonPath":".spec.driver"},{"name":"VolumeSnapshotClass","type":"string","description":"Name of the VolumeSnapshotClass to which this snapshot belongs.","jsonPath":".spec.volumeSnapshotClassName"},{"name":"VolumeSnapshot","type":"string","description":"Name of the VolumeSnapshot object to which this VolumeSnapshotContent object is bound.","jsonPath":".spec.volumeSnapshotRef.name"},{"name":"VolumeSnapshotNamespace","type":"string","description":"Namespace of the VolumeSnapshot object to which this VolumeSnapshotContent object is bound.","jsonPath":".spec.volumeSnapshotRef.namespace"},{"name":"Age","type":"date","jsonPath":".metadata.creationTimestamp"}]},{"name":"v1beta1","served":false,"storage":false,"deprecated":true,"deprecationWarning":"snapshot.storage.k8s.io/v1beta1 VolumeSnapshotContent is deprecated; use snapshot.storage.k8s.io/v1 VolumeSnapshotContent","schema":{"openAPIV3Schema":{"description":"VolumeSnapshotContent represents the actual \"on-disk\" snapshot object in the underlying storage system","type":"object","required":["spec"],"properties":{"apiVersion":{"description":"APIVersion defines the versioned schema of this representation of an object. Servers should convert recognized schemas to the latest internal value, and may reject unrecognized values. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#resources","type":"string"},"kind":{"description":"Kind is a string value representing the REST resource this object represents. Servers may infer this from the endpoint the client submits requests to. Cannot be updated. In CamelCase. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#types-kinds","type":"string"},"spec":{"description":"spec defines properties of a VolumeSnapshotContent created by the underlying storage system. Required.","type":"object","required":["deletionPolicy","driver","source","volumeSnapshotRef"],"properties":{"deletionPolicy":{"description":"deletionPolicy determines whether this VolumeSnapshotContent and its physical snapshot on the underlying storage system should be deleted when its bound VolumeSnapshot is deleted. Supported values are \"Retain\" and \"Delete\". \"Retain\" means that the VolumeSnapshotContent and its physical snapshot on underlying storage system are kept. \"Delete\" means that the VolumeSnapshotContent and its physical snapshot on underlying storage system are deleted. For dynamically provisioned snapshots, this field will automatically be filled in by the CSI snapshotter sidecar with the \"DeletionPolicy\" field defined in the corresponding VolumeSnapshotClass. For pre-existing snapshots, users MUST specify this field when creating the VolumeSnapshotContent object. Required.","type":"string","enum":["Delete","Retain"]},"driver":{"description":"driver is the name of the CSI driver used to create the physical snapshot on the underlying storage system. This MUST be the same as the name returned by the CSI GetPluginName() call for that driver. Required.","type":"string"},"source":{"description":"source specifies whether the snapshot is (or should be) dynamically provisioned or already exists, and just requires a Kubernetes object representation. This field is immutable after creation. Required.","type":"object","properties":{"snapshotHandle":{"description":"snapshotHandle specifies the CSI \"snapshot_id\" of a pre-existing snapshot on the underlying storage system for which a Kubernetes object representation was (or should be) created. This field is immutable.","type":"string"},"volumeHandle":{"description":"volumeHandle specifies the CSI \"volume_id\" of the volume from which a snapshot should be dynamically taken from. This field is immutable.","type":"string"}}},"volumeSnapshotClassName":{"description":"name of the VolumeSnapshotClass from which this snapshot was (or will be) created. Note that after provisioning, the VolumeSnapshotClass may be deleted or recreated with different set of values, and as such, should not be referenced post-snapshot creation.","type":"string"},"volumeSnapshotRef":{"description":"volumeSnapshotRef specifies the VolumeSnapshot object to which this VolumeSnapshotContent object is bound. VolumeSnapshot.Spec.VolumeSnapshotContentName field must reference to this VolumeSnapshotContent's name for the bidirectional binding to be valid. For a pre-existing VolumeSnapshotContent object, name and namespace of the VolumeSnapshot object MUST be provided for binding to happen. This field is immutable after creation. Required.","type":"object","properties":{"apiVersion":{"description":"API version of the referent.","type":"string"},"fieldPath":{"description":"If referring to a piece of an object instead of an entire object, this string should contain a valid JSON/Go field access statement, such as desiredState.manifest.containers[2]. For example, if the object reference is to a container within a pod, this would take on a value like: \"spec.containers{name}\" (where \"name\" refers to the name of the container that triggered the event) or if no container name is specified \"spec.containers[2]\" (container with index 2 in this pod). This syntax is chosen only to have some well-defined way of referencing a part of an object. TODO: this design is not final and this field is subject to change in the future.","type":"string"},"kind":{"description":"Kind of the referent. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#types-kinds","type":"string"},"name":{"description":"Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names","type":"string"},"namespace":{"description":"Namespace of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/namespaces/","type":"string"},"resourceVersion":{"description":"Specific resourceVersion to which this reference is made, if any. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#concurrency-control-and-consistency","type":"string"},"uid":{"description":"UID of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#uids","type":"string"}}}}},"status":{"description":"status represents the current information of a snapshot.","type":"object","properties":{"creationTime":{"description":"creationTime is the timestamp when the point-in-time snapshot is taken by the underlying storage system. In dynamic snapshot creation case, this field will be filled in by the CSI snapshotter sidecar with the \"creation_time\" value returned from CSI \"CreateSnapshot\" gRPC call. For a pre-existing snapshot, this field will be filled with the \"creation_time\" value returned from the CSI \"ListSnapshots\" gRPC call if the driver supports it. If not specified, it indicates the creation time is unknown. The format of this field is a Unix nanoseconds time encoded as an int64. On Unix, the command `date +%s%N` returns the current time in nanoseconds since 1970-01-01 00:00:00 UTC.","type":"integer","format":"int64"},"error":{"description":"error is the last observed error during snapshot creation, if any. Upon success after retry, this error field will be cleared.","type":"object","properties":{"message":{"description":"message is a string detailing the encountered error during snapshot creation if specified. NOTE: message may be logged, and it should not contain sensitive information.","type":"string"},"time":{"description":"time is the timestamp when the error was encountered.","type":"string","format":"date-time"}}},"readyToUse":{"description":"readyToUse indicates if a snapshot is ready to be used to restore a volume. In dynamic snapshot creation case, this field will be filled in by the CSI snapshotter sidecar with the \"ready_to_use\" value returned from CSI \"CreateSnapshot\" gRPC call. For a pre-existing snapshot, this field will be filled with the \"ready_to_use\" value returned from the CSI \"ListSnapshots\" gRPC call if the driver supports it, otherwise, this field will be set to \"True\". If not specified, it means the readiness of a snapshot is unknown.","type":"boolean"},"restoreSize":{"description":"restoreSize represents the complete size of the snapshot in bytes. In dynamic snapshot creation case, this field will be filled in by the CSI snapshotter sidecar with the \"size_bytes\" value returned from CSI \"CreateSnapshot\" gRPC call. For a pre-existing snapshot, this field will be filled with the \"size_bytes\" value returned from the CSI \"ListSnapshots\" gRPC call if the driver supports it. When restoring a volume from this snapshot, the size of the volume MUST NOT be smaller than the restoreSize if it is specified, otherwise the restoration will fail. If not specified, it indicates that the size is unknown.","type":"integer","format":"int64","minimum":0},"snapshotHandle":{"description":"snapshotHandle is the CSI \"snapshot_id\" of a snapshot on the underlying storage system. If not specified, it indicates that dynamic snapshot creation has either failed or it is still in progress.","type":"string"}}}}}},"subresources":{"status":{}},"additionalPrinterColumns":[{"name":"ReadyToUse","type":"boolean","description":"Indicates if the snapshot is ready to be used to restore a volume.","jsonPath":".status.readyToUse"},{"name":"RestoreSize","type":"integer","description":"Represents the complete size of the snapshot in bytes","jsonPath":".status.restoreSize"},{"name":"DeletionPolicy","type":"string","description":"Determines whether this VolumeSnapshotContent and its physical snapshot on the underlying storage system should be deleted when its bound VolumeSnapshot is deleted.","jsonPath":".spec.deletionPolicy"},{"name":"Driver","type":"string","description":"Name of the CSI driver used to create the physical snapshot on the underlying storage system.","jsonPath":".spec.driver"},{"name":"VolumeSnapshotClass","type":"string","description":"Name of the VolumeSnapshotClass to which this snapshot belongs.","jsonPath":".spec.volumeSnapshotClassName"},{"name":"VolumeSnapshot","type":"string","description":"Name of the VolumeSnapshot object to which this VolumeSnapshotContent object is bound.","jsonPath":".spec.volumeSnapshotRef.name"},{"name":"VolumeSnapshotNamespace","type":"string","description":"Namespace of the VolumeSnapshot object to which this VolumeSnapshotContent object is bound.","jsonPath":".spec.volumeSnapshotRef.namespace"},{"name":"Age","type":"date","jsonPath":".metadata.creationTimestamp"}]}],"conversion":{"strategy":"None"}},"status":{"conditions":[{"type":"KubernetesAPIApprovalPolicyConformant","status":"True","lastTransitionTime":"2025-12-04T11:37:44Z","reason":"ApprovedAnnotation","message":"approved in https://github.com/kubernetes-csi/external-snapshotter/pull/955"},{"type":"NamesAccepted","status":"True","lastTransitionTime":"2025-12-04T11:37:44Z","reason":"NoConflicts","message":"no conflicts found"},{"type":"Established","status":"True","lastTransitionTime":"2025-12-04T11:37:44Z","reason":"InitialNamesAccepted","message":"the initial names have been accepted"}],"acceptedNames":{"plural":"volumesnapshotcontents","singular":"volumesnapshotcontent","shortNames":["vsc","vscs"],"kind":"VolumeSnapshotContent","listKind":"VolumeSnapshotContentList"},"storedVersions":["v1"]}}config/crd/volumesnapshots.snapshot.storage.k8s.io.json0000640000000000000000000004563015114274061022046 0ustar0000000000000000{"metadata":{"name":"volumesnapshots.snapshot.storage.k8s.io","uid":"7f61848b-c1fc-4e18-a2cf-c7f775d4810e","resourceVersion":"3936","generation":1,"creationTimestamp":"2025-12-04T11:37:44Z","annotations":{"api-approved.kubernetes.io":"https://github.com/kubernetes-csi/external-snapshotter/pull/814","controller-gen.kubebuilder.io/version":"v0.15.0"}},"spec":{"group":"snapshot.storage.k8s.io","names":{"plural":"volumesnapshots","singular":"volumesnapshot","shortNames":["vs"],"kind":"VolumeSnapshot","listKind":"VolumeSnapshotList"},"scope":"Namespaced","versions":[{"name":"v1","served":true,"storage":true,"schema":{"openAPIV3Schema":{"description":"VolumeSnapshot is a user's request for either creating a point-in-time\nsnapshot of a persistent volume, or binding to a pre-existing snapshot.","type":"object","required":["spec"],"properties":{"apiVersion":{"description":"APIVersion defines the versioned schema of this representation of an object.\nServers should convert recognized schemas to the latest internal value, and\nmay reject unrecognized values.\nMore info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#resources","type":"string"},"kind":{"description":"Kind is a string value representing the REST resource this object represents.\nServers may infer this from the endpoint the client submits requests to.\nCannot be updated.\nIn CamelCase.\nMore info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#types-kinds","type":"string"},"metadata":{"type":"object"},"spec":{"description":"spec defines the desired characteristics of a snapshot requested by a user.\nMore info: https://kubernetes.io/docs/concepts/storage/volume-snapshots#volumesnapshots\nRequired.","type":"object","required":["source"],"properties":{"source":{"description":"source specifies where a snapshot will be created from.\nThis field is immutable after creation.\nRequired.","type":"object","properties":{"persistentVolumeClaimName":{"description":"persistentVolumeClaimName specifies the name of the PersistentVolumeClaim\nobject representing the volume from which a snapshot should be created.\nThis PVC is assumed to be in the same namespace as the VolumeSnapshot\nobject.\nThis field should be set if the snapshot does not exists, and needs to be\ncreated.\nThis field is immutable.","type":"string","x-kubernetes-validations":[{"rule":"self == oldSelf","message":"persistentVolumeClaimName is immutable"}]},"volumeSnapshotContentName":{"description":"volumeSnapshotContentName specifies the name of a pre-existing VolumeSnapshotContent\nobject representing an existing volume snapshot.\nThis field should be set if the snapshot already exists and only needs a representation in Kubernetes.\nThis field is immutable.","type":"string","x-kubernetes-validations":[{"rule":"self == oldSelf","message":"volumeSnapshotContentName is immutable"}]}},"x-kubernetes-validations":[{"rule":"!has(oldSelf.persistentVolumeClaimName) || has(self.persistentVolumeClaimName)","message":"persistentVolumeClaimName is required once set"},{"rule":"!has(oldSelf.volumeSnapshotContentName) || has(self.volumeSnapshotContentName)","message":"volumeSnapshotContentName is required once set"},{"rule":"(has(self.volumeSnapshotContentName) \u0026\u0026 !has(self.persistentVolumeClaimName)) || (!has(self.volumeSnapshotContentName) \u0026\u0026 has(self.persistentVolumeClaimName))","message":"exactly one of volumeSnapshotContentName and persistentVolumeClaimName must be set"}]},"volumeSnapshotClassName":{"description":"VolumeSnapshotClassName is the name of the VolumeSnapshotClass\nrequested by the VolumeSnapshot.\nVolumeSnapshotClassName may be left nil to indicate that the default\nSnapshotClass should be used.\nA given cluster may have multiple default Volume SnapshotClasses: one\ndefault per CSI Driver. If a VolumeSnapshot does not specify a SnapshotClass,\nVolumeSnapshotSource will be checked to figure out what the associated\nCSI Driver is, and the default VolumeSnapshotClass associated with that\nCSI Driver will be used. If more than one VolumeSnapshotClass exist for\na given CSI Driver and more than one have been marked as default,\nCreateSnapshot will fail and generate an event.\nEmpty string is not allowed for this field.","type":"string","x-kubernetes-validations":[{"rule":"size(self) \u003e 0","message":"volumeSnapshotClassName must not be the empty string when set"}]}}},"status":{"description":"status represents the current information of a snapshot.\nConsumers must verify binding between VolumeSnapshot and\nVolumeSnapshotContent objects is successful (by validating that both\nVolumeSnapshot and VolumeSnapshotContent point at each other) before\nusing this object.","type":"object","properties":{"boundVolumeSnapshotContentName":{"description":"boundVolumeSnapshotContentName is the name of the VolumeSnapshotContent\nobject to which this VolumeSnapshot object intends to bind to.\nIf not specified, it indicates that the VolumeSnapshot object has not been\nsuccessfully bound to a VolumeSnapshotContent object yet.\nNOTE: To avoid possible security issues, consumers must verify binding between\nVolumeSnapshot and VolumeSnapshotContent objects is successful (by validating that\nboth VolumeSnapshot and VolumeSnapshotContent point at each other) before using\nthis object.","type":"string"},"creationTime":{"description":"creationTime is the timestamp when the point-in-time snapshot is taken\nby the underlying storage system.\nIn dynamic snapshot creation case, this field will be filled in by the\nsnapshot controller with the \"creation_time\" value returned from CSI\n\"CreateSnapshot\" gRPC call.\nFor a pre-existing snapshot, this field will be filled with the \"creation_time\"\nvalue returned from the CSI \"ListSnapshots\" gRPC call if the driver supports it.\nIf not specified, it may indicate that the creation time of the snapshot is unknown.","type":"string","format":"date-time"},"error":{"description":"error is the last observed error during snapshot creation, if any.\nThis field could be helpful to upper level controllers(i.e., application controller)\nto decide whether they should continue on waiting for the snapshot to be created\nbased on the type of error reported.\nThe snapshot controller will keep retrying when an error occurs during the\nsnapshot creation. Upon success, this error field will be cleared.","type":"object","properties":{"message":{"description":"message is a string detailing the encountered error during snapshot\ncreation if specified.\nNOTE: message may be logged, and it should not contain sensitive\ninformation.","type":"string"},"time":{"description":"time is the timestamp when the error was encountered.","type":"string","format":"date-time"}}},"readyToUse":{"description":"readyToUse indicates if the snapshot is ready to be used to restore a volume.\nIn dynamic snapshot creation case, this field will be filled in by the\nsnapshot controller with the \"ready_to_use\" value returned from CSI\n\"CreateSnapshot\" gRPC call.\nFor a pre-existing snapshot, this field will be filled with the \"ready_to_use\"\nvalue returned from the CSI \"ListSnapshots\" gRPC call if the driver supports it,\notherwise, this field will be set to \"True\".\nIf not specified, it means the readiness of a snapshot is unknown.","type":"boolean"},"restoreSize":{"description":"restoreSize represents the minimum size of volume required to create a volume\nfrom this snapshot.\nIn dynamic snapshot creation case, this field will be filled in by the\nsnapshot controller with the \"size_bytes\" value returned from CSI\n\"CreateSnapshot\" gRPC call.\nFor a pre-existing snapshot, this field will be filled with the \"size_bytes\"\nvalue returned from the CSI \"ListSnapshots\" gRPC call if the driver supports it.\nWhen restoring a volume from this snapshot, the size of the volume MUST NOT\nbe smaller than the restoreSize if it is specified, otherwise the restoration will fail.\nIf not specified, it indicates that the size is unknown.","type":"string","pattern":"^(\\+|-)?(([0-9]+(\\.[0-9]*)?)|(\\.[0-9]+))(([KMGTPE]i)|[numkMGTPE]|([eE](\\+|-)?(([0-9]+(\\.[0-9]*)?)|(\\.[0-9]+))))?$","x-kubernetes-int-or-string":true},"volumeGroupSnapshotName":{"description":"VolumeGroupSnapshotName is the name of the VolumeGroupSnapshot of which this\nVolumeSnapshot is a part of.","type":"string"}}}}}},"subresources":{"status":{}},"additionalPrinterColumns":[{"name":"ReadyToUse","type":"boolean","description":"Indicates if the snapshot is ready to be used to restore a volume.","jsonPath":".status.readyToUse"},{"name":"SourcePVC","type":"string","description":"If a new snapshot needs to be created, this contains the name of the source PVC from which this snapshot was (or will be) created.","jsonPath":".spec.source.persistentVolumeClaimName"},{"name":"SourceSnapshotContent","type":"string","description":"If a snapshot already exists, this contains the name of the existing VolumeSnapshotContent object representing the existing snapshot.","jsonPath":".spec.source.volumeSnapshotContentName"},{"name":"RestoreSize","type":"string","description":"Represents the minimum size of volume required to rehydrate from this snapshot.","jsonPath":".status.restoreSize"},{"name":"SnapshotClass","type":"string","description":"The name of the VolumeSnapshotClass requested by the VolumeSnapshot.","jsonPath":".spec.volumeSnapshotClassName"},{"name":"SnapshotContent","type":"string","description":"Name of the VolumeSnapshotContent object to which the VolumeSnapshot object intends to bind to. Please note that verification of binding actually requires checking both VolumeSnapshot and VolumeSnapshotContent to ensure both are pointing at each other. Binding MUST be verified prior to usage of this object.","jsonPath":".status.boundVolumeSnapshotContentName"},{"name":"CreationTime","type":"date","description":"Timestamp when the point-in-time snapshot was taken by the underlying storage system.","jsonPath":".status.creationTime"},{"name":"Age","type":"date","jsonPath":".metadata.creationTimestamp"}]},{"name":"v1beta1","served":false,"storage":false,"deprecated":true,"deprecationWarning":"snapshot.storage.k8s.io/v1beta1 VolumeSnapshot is deprecated; use snapshot.storage.k8s.io/v1 VolumeSnapshot","schema":{"openAPIV3Schema":{"description":"VolumeSnapshot is a user's request for either creating a point-in-time snapshot of a persistent volume, or binding to a pre-existing snapshot.","type":"object","required":["spec"],"properties":{"apiVersion":{"description":"APIVersion defines the versioned schema of this representation of an object. Servers should convert recognized schemas to the latest internal value, and may reject unrecognized values. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#resources","type":"string"},"kind":{"description":"Kind is a string value representing the REST resource this object represents. Servers may infer this from the endpoint the client submits requests to. Cannot be updated. In CamelCase. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#types-kinds","type":"string"},"spec":{"description":"spec defines the desired characteristics of a snapshot requested by a user. More info: https://kubernetes.io/docs/concepts/storage/volume-snapshots#volumesnapshots Required.","type":"object","required":["source"],"properties":{"source":{"description":"source specifies where a snapshot will be created from. This field is immutable after creation. Required.","type":"object","properties":{"persistentVolumeClaimName":{"description":"persistentVolumeClaimName specifies the name of the PersistentVolumeClaim object representing the volume from which a snapshot should be created. This PVC is assumed to be in the same namespace as the VolumeSnapshot object. This field should be set if the snapshot does not exists, and needs to be created. This field is immutable.","type":"string"},"volumeSnapshotContentName":{"description":"volumeSnapshotContentName specifies the name of a pre-existing VolumeSnapshotContent object representing an existing volume snapshot. This field should be set if the snapshot already exists and only needs a representation in Kubernetes. This field is immutable.","type":"string"}}},"volumeSnapshotClassName":{"description":"VolumeSnapshotClassName is the name of the VolumeSnapshotClass requested by the VolumeSnapshot. VolumeSnapshotClassName may be left nil to indicate that the default SnapshotClass should be used. A given cluster may have multiple default Volume SnapshotClasses: one default per CSI Driver. If a VolumeSnapshot does not specify a SnapshotClass, VolumeSnapshotSource will be checked to figure out what the associated CSI Driver is, and the default VolumeSnapshotClass associated with that CSI Driver will be used. If more than one VolumeSnapshotClass exist for a given CSI Driver and more than one have been marked as default, CreateSnapshot will fail and generate an event. Empty string is not allowed for this field.","type":"string"}}},"status":{"description":"status represents the current information of a snapshot. Consumers must verify binding between VolumeSnapshot and VolumeSnapshotContent objects is successful (by validating that both VolumeSnapshot and VolumeSnapshotContent point at each other) before using this object.","type":"object","properties":{"boundVolumeSnapshotContentName":{"description":"boundVolumeSnapshotContentName is the name of the VolumeSnapshotContent object to which this VolumeSnapshot object intends to bind to. If not specified, it indicates that the VolumeSnapshot object has not been successfully bound to a VolumeSnapshotContent object yet. NOTE: To avoid possible security issues, consumers must verify binding between VolumeSnapshot and VolumeSnapshotContent objects is successful (by validating that both VolumeSnapshot and VolumeSnapshotContent point at each other) before using this object.","type":"string"},"creationTime":{"description":"creationTime is the timestamp when the point-in-time snapshot is taken by the underlying storage system. In dynamic snapshot creation case, this field will be filled in by the snapshot controller with the \"creation_time\" value returned from CSI \"CreateSnapshot\" gRPC call. For a pre-existing snapshot, this field will be filled with the \"creation_time\" value returned from the CSI \"ListSnapshots\" gRPC call if the driver supports it. If not specified, it may indicate that the creation time of the snapshot is unknown.","type":"string","format":"date-time"},"error":{"description":"error is the last observed error during snapshot creation, if any. This field could be helpful to upper level controllers(i.e., application controller) to decide whether they should continue on waiting for the snapshot to be created based on the type of error reported. The snapshot controller will keep retrying when an error occurs during the snapshot creation. Upon success, this error field will be cleared.","type":"object","properties":{"message":{"description":"message is a string detailing the encountered error during snapshot creation if specified. NOTE: message may be logged, and it should not contain sensitive information.","type":"string"},"time":{"description":"time is the timestamp when the error was encountered.","type":"string","format":"date-time"}}},"readyToUse":{"description":"readyToUse indicates if the snapshot is ready to be used to restore a volume. In dynamic snapshot creation case, this field will be filled in by the snapshot controller with the \"ready_to_use\" value returned from CSI \"CreateSnapshot\" gRPC call. For a pre-existing snapshot, this field will be filled with the \"ready_to_use\" value returned from the CSI \"ListSnapshots\" gRPC call if the driver supports it, otherwise, this field will be set to \"True\". If not specified, it means the readiness of a snapshot is unknown.","type":"boolean"},"restoreSize":{"description":"restoreSize represents the minimum size of volume required to create a volume from this snapshot. In dynamic snapshot creation case, this field will be filled in by the snapshot controller with the \"size_bytes\" value returned from CSI \"CreateSnapshot\" gRPC call. For a pre-existing snapshot, this field will be filled with the \"size_bytes\" value returned from the CSI \"ListSnapshots\" gRPC call if the driver supports it. When restoring a volume from this snapshot, the size of the volume MUST NOT be smaller than the restoreSize if it is specified, otherwise the restoration will fail. If not specified, it indicates that the size is unknown.","type":"string","pattern":"^(\\+|-)?(([0-9]+(\\.[0-9]*)?)|(\\.[0-9]+))(([KMGTPE]i)|[numkMGTPE]|([eE](\\+|-)?(([0-9]+(\\.[0-9]*)?)|(\\.[0-9]+))))?$","x-kubernetes-int-or-string":true}}}}}},"subresources":{"status":{}},"additionalPrinterColumns":[{"name":"ReadyToUse","type":"boolean","description":"Indicates if the snapshot is ready to be used to restore a volume.","jsonPath":".status.readyToUse"},{"name":"SourcePVC","type":"string","description":"If a new snapshot needs to be created, this contains the name of the source PVC from which this snapshot was (or will be) created.","jsonPath":".spec.source.persistentVolumeClaimName"},{"name":"SourceSnapshotContent","type":"string","description":"If a snapshot already exists, this contains the name of the existing VolumeSnapshotContent object representing the existing snapshot.","jsonPath":".spec.source.volumeSnapshotContentName"},{"name":"RestoreSize","type":"string","description":"Represents the minimum size of volume required to rehydrate from this snapshot.","jsonPath":".status.restoreSize"},{"name":"SnapshotClass","type":"string","description":"The name of the VolumeSnapshotClass requested by the VolumeSnapshot.","jsonPath":".spec.volumeSnapshotClassName"},{"name":"SnapshotContent","type":"string","description":"Name of the VolumeSnapshotContent object to which the VolumeSnapshot object intends to bind to. Please note that verification of binding actually requires checking both VolumeSnapshot and VolumeSnapshotContent to ensure both are pointing at each other. Binding MUST be verified prior to usage of this object.","jsonPath":".status.boundVolumeSnapshotContentName"},{"name":"CreationTime","type":"date","description":"Timestamp when the point-in-time snapshot was taken by the underlying storage system.","jsonPath":".status.creationTime"},{"name":"Age","type":"date","jsonPath":".metadata.creationTimestamp"}]}],"conversion":{"strategy":"None"}},"status":{"conditions":[{"type":"NamesAccepted","status":"True","lastTransitionTime":"2025-12-04T11:37:44Z","reason":"NoConflicts","message":"no conflicts found"},{"type":"Established","status":"True","lastTransitionTime":"2025-12-04T11:37:44Z","reason":"InitialNamesAccepted","message":"the initial names have been accepted"},{"type":"KubernetesAPIApprovalPolicyConformant","status":"True","lastTransitionTime":"2025-12-04T11:37:44Z","reason":"ApprovedAnnotation","message":"approved in https://github.com/kubernetes-csi/external-snapshotter/pull/814"}],"acceptedNames":{"plural":"volumesnapshots","singular":"volumesnapshot","shortNames":["vs"],"kind":"VolumeSnapshot","listKind":"VolumeSnapshotList"},"storedVersions":["v1"]}}config/namespaces_with_overlapping_uids.json0000640000000000000000000000000215114274061020175 0ustar0000000000000000[]cluster-scoped-resources/rbac.authorization.k8s.io/clusterroles/edit.json0000640000000000000000000002215615114274061025501 0ustar0000000000000000{"metadata":{"name":"edit","uid":"2559835b-126d-4c15-8021-eaf417206e9c","resourceVersion":"7554","creationTimestamp":"2025-12-04T11:29:40Z","labels":{"kubernetes.io/bootstrapping":"rbac-defaults","rbac.authorization.k8s.io/aggregate-to-admin":"true"},"annotations":{"rbac.authorization.kubernetes.io/autoupdate":"true"}},"rules":[{"verbs":["create","update","patch","delete"],"apiGroups":["operators.coreos.com"],"resources":["subscriptions"]},{"verbs":["delete"],"apiGroups":["operators.coreos.com"],"resources":["clusterserviceversions","catalogsources","installplans","subscriptions"]},{"verbs":["get","list","watch"],"apiGroups":["operators.coreos.com"],"resources":["clusterserviceversions","catalogsources","installplans","subscriptions","operatorgroups"]},{"verbs":["get","list","watch"],"apiGroups":["packages.operators.coreos.com"],"resources":["packagemanifests","packagemanifests/icon"]},{"verbs":["create","update","patch","delete"],"apiGroups":["k8s.ovn.org"],"resources":["userdefinednetworks"]},{"verbs":["create","delete","deletecollection","get","list","patch","update","watch"],"apiGroups":[""],"resources":["secrets","serviceaccounts"]},{"verbs":["create","delete","deletecollection","get","list","patch","update","watch"],"apiGroups":["","image.openshift.io"],"resources":["imagestreamimages","imagestreammappings","imagestreams","imagestreams/secrets","imagestreamtags","imagetags"]},{"verbs":["create"],"apiGroups":["","image.openshift.io"],"resources":["imagestreamimports"]},{"verbs":["get","update"],"apiGroups":["","image.openshift.io"],"resources":["imagestreams/layers"]},{"verbs":["get"],"apiGroups":[""],"resources":["namespaces"]},{"verbs":["get"],"apiGroups":["","project.openshift.io"],"resources":["projects"]},{"verbs":["get","list","watch"],"apiGroups":[""],"resources":["pods/attach","pods/exec","pods/portforward","pods/proxy","secrets","services/proxy"]},{"verbs":["impersonate"],"apiGroups":[""],"resources":["serviceaccounts"]},{"verbs":["create","delete","deletecollection","patch","update"],"apiGroups":[""],"resources":["pods","pods/attach","pods/exec","pods/portforward","pods/proxy"]},{"verbs":["create"],"apiGroups":[""],"resources":["pods/eviction"]},{"verbs":["create","delete","deletecollection","patch","update"],"apiGroups":[""],"resources":["configmaps","endpoints","events","persistentvolumeclaims","replicationcontrollers","replicationcontrollers/scale","secrets","serviceaccounts","services","services/proxy"]},{"verbs":["create"],"apiGroups":[""],"resources":["serviceaccounts/token"]},{"verbs":["create","delete","deletecollection","patch","update"],"apiGroups":["apps"],"resources":["daemonsets","deployments","deployments/rollback","deployments/scale","replicasets","replicasets/scale","statefulsets","statefulsets/scale"]},{"verbs":["create","delete","deletecollection","patch","update"],"apiGroups":["autoscaling"],"resources":["horizontalpodautoscalers"]},{"verbs":["create","delete","deletecollection","patch","update"],"apiGroups":["batch"],"resources":["cronjobs","jobs"]},{"verbs":["create","delete","deletecollection","patch","update"],"apiGroups":["extensions"],"resources":["daemonsets","deployments","deployments/rollback","deployments/scale","ingresses","networkpolicies","replicasets","replicasets/scale","replicationcontrollers/scale"]},{"verbs":["create","delete","deletecollection","patch","update"],"apiGroups":["policy"],"resources":["poddisruptionbudgets"]},{"verbs":["create","delete","deletecollection","patch","update"],"apiGroups":["networking.k8s.io"],"resources":["ingresses","networkpolicies"]},{"verbs":["create","delete","deletecollection","get","list","patch","update","watch"],"apiGroups":["coordination.k8s.io"],"resources":["leases"]},{"verbs":["create"],"apiGroups":["","image.openshift.io"],"resources":["imagestreams"]},{"verbs":["update"],"apiGroups":["","build.openshift.io"],"resources":["builds/details"]},{"verbs":["get"],"apiGroups":["","build.openshift.io"],"resources":["builds"]},{"verbs":["get","list","watch","create","update","patch","delete","deletecollection"],"apiGroups":["snapshot.storage.k8s.io"],"resources":["volumesnapshots"]},{"verbs":["create","delete","deletecollection","get","list","patch","update","watch"],"apiGroups":["","build.openshift.io"],"resources":["buildconfigs","buildconfigs/webhooks","builds"]},{"verbs":["get","list","watch"],"apiGroups":["","build.openshift.io"],"resources":["builds/log"]},{"verbs":["create"],"apiGroups":["","build.openshift.io"],"resources":["buildconfigs/instantiate","buildconfigs/instantiatebinary","builds/clone"]},{"verbs":["edit","view"],"apiGroups":["build.openshift.io"],"resources":["jenkins"]},{"verbs":["create","delete","deletecollection","get","list","patch","update","watch"],"apiGroups":["","apps.openshift.io"],"resources":["deploymentconfigs","deploymentconfigs/scale"]},{"verbs":["create"],"apiGroups":["","apps.openshift.io"],"resources":["deploymentconfigrollbacks","deploymentconfigs/instantiate","deploymentconfigs/rollback"]},{"verbs":["get","list","watch"],"apiGroups":["","apps.openshift.io"],"resources":["deploymentconfigs/log","deploymentconfigs/status"]},{"verbs":["get","list","watch"],"apiGroups":["","image.openshift.io"],"resources":["imagestreams/status"]},{"verbs":["get","list","watch"],"apiGroups":["","quota.openshift.io"],"resources":["appliedclusterresourcequotas"]},{"verbs":["create","delete","deletecollection","get","list","patch","update","watch"],"apiGroups":["","route.openshift.io"],"resources":["routes"]},{"verbs":["create"],"apiGroups":["","route.openshift.io"],"resources":["routes/custom-host"]},{"verbs":["get","list","watch"],"apiGroups":["","route.openshift.io"],"resources":["routes/status"]},{"verbs":["create","delete","deletecollection","get","list","patch","update","watch"],"apiGroups":["","template.openshift.io"],"resources":["processedtemplates","templateconfigs","templateinstances","templates"]},{"verbs":["create","delete","deletecollection","get","list","patch","update","watch"],"apiGroups":["networking.k8s.io"],"resources":["networkpolicies"]},{"verbs":["create","delete","deletecollection","get","list","patch","update","watch"],"apiGroups":["","build.openshift.io"],"resources":["buildlogs"]},{"verbs":["get","list","watch"],"apiGroups":[""],"resources":["resourcequotausages"]},{"verbs":["get","list","watch"],"apiGroups":["k8s.ovn.org"],"resources":["userdefinednetworks"]},{"verbs":["get","list","watch"],"apiGroups":["","image.openshift.io"],"resources":["imagestreamimages","imagestreammappings","imagestreams","imagestreamtags","imagetags"]},{"verbs":["get"],"apiGroups":["","image.openshift.io"],"resources":["imagestreams/layers"]},{"verbs":["get","list","watch"],"apiGroups":[""],"resources":["configmaps","endpoints","persistentvolumeclaims","persistentvolumeclaims/status","pods","replicationcontrollers","replicationcontrollers/scale","serviceaccounts","services","services/status"]},{"verbs":["get","list","watch"],"apiGroups":[""],"resources":["bindings","events","limitranges","namespaces/status","pods/log","pods/status","replicationcontrollers/status","resourcequotas","resourcequotas/status"]},{"verbs":["get","list","watch"],"apiGroups":[""],"resources":["namespaces"]},{"verbs":["get","list","watch"],"apiGroups":["discovery.k8s.io"],"resources":["endpointslices"]},{"verbs":["get","list","watch"],"apiGroups":["apps"],"resources":["controllerrevisions","daemonsets","daemonsets/status","deployments","deployments/scale","deployments/status","replicasets","replicasets/scale","replicasets/status","statefulsets","statefulsets/scale","statefulsets/status"]},{"verbs":["get","list","watch"],"apiGroups":["autoscaling"],"resources":["horizontalpodautoscalers","horizontalpodautoscalers/status"]},{"verbs":["get","list","watch"],"apiGroups":["batch"],"resources":["cronjobs","cronjobs/status","jobs","jobs/status"]},{"verbs":["get","list","watch"],"apiGroups":["extensions"],"resources":["daemonsets","daemonsets/status","deployments","deployments/scale","deployments/status","ingresses","ingresses/status","networkpolicies","replicasets","replicasets/scale","replicasets/status","replicationcontrollers/scale"]},{"verbs":["get","list","watch"],"apiGroups":["policy"],"resources":["poddisruptionbudgets","poddisruptionbudgets/status"]},{"verbs":["get","list","watch"],"apiGroups":["networking.k8s.io"],"resources":["ingresses","ingresses/status","networkpolicies"]},{"verbs":["get","list","watch"],"apiGroups":["snapshot.storage.k8s.io"],"resources":["volumesnapshots"]},{"verbs":["get","list","watch"],"apiGroups":["","build.openshift.io"],"resources":["buildconfigs","buildconfigs/webhooks","builds"]},{"verbs":["view"],"apiGroups":["build.openshift.io"],"resources":["jenkins"]},{"verbs":["get","list","watch"],"apiGroups":["","apps.openshift.io"],"resources":["deploymentconfigs","deploymentconfigs/scale"]},{"verbs":["get","list","watch"],"apiGroups":["","route.openshift.io"],"resources":["routes"]},{"verbs":["get","list","watch"],"apiGroups":["","template.openshift.io"],"resources":["processedtemplates","templateconfigs","templateinstances","templates"]},{"verbs":["get","list","watch"],"apiGroups":["","build.openshift.io"],"resources":["buildlogs"]}],"aggregationRule":{"clusterRoleSelectors":[{"matchLabels":{"rbac.authorization.k8s.io/aggregate-to-edit":"true"}}]}}cluster-scoped-resources/rbac.authorization.k8s.io/clusterroles/admin.json0000640000000000000000000002504315114274061025642 0ustar0000000000000000{"metadata":{"name":"admin","uid":"d3f91e04-d191-4878-b38f-3e61baa4e5aa","resourceVersion":"7563","creationTimestamp":"2025-12-04T11:29:40Z","labels":{"kubernetes.io/bootstrapping":"rbac-defaults"},"annotations":{"rbac.authorization.kubernetes.io/autoupdate":"true"}},"rules":[{"verbs":["create","update","patch","delete"],"apiGroups":["operators.coreos.com"],"resources":["subscriptions"]},{"verbs":["delete"],"apiGroups":["operators.coreos.com"],"resources":["clusterserviceversions","catalogsources","installplans","subscriptions"]},{"verbs":["get","list","watch"],"apiGroups":["operators.coreos.com"],"resources":["clusterserviceversions","catalogsources","installplans","subscriptions","operatorgroups"]},{"verbs":["get","list","watch"],"apiGroups":["packages.operators.coreos.com"],"resources":["packagemanifests","packagemanifests/icon"]},{"verbs":["create","update","patch","delete"],"apiGroups":["k8s.ovn.org"],"resources":["userdefinednetworks"]},{"verbs":["create","delete","deletecollection","get","list","patch","update","watch"],"apiGroups":[""],"resources":["secrets","serviceaccounts"]},{"verbs":["create","delete","deletecollection","get","list","patch","update","watch"],"apiGroups":["","image.openshift.io"],"resources":["imagestreamimages","imagestreammappings","imagestreams","imagestreams/secrets","imagestreamtags","imagetags"]},{"verbs":["create"],"apiGroups":["","image.openshift.io"],"resources":["imagestreamimports"]},{"verbs":["get","update"],"apiGroups":["","image.openshift.io"],"resources":["imagestreams/layers"]},{"verbs":["get"],"apiGroups":[""],"resources":["namespaces"]},{"verbs":["get"],"apiGroups":["","project.openshift.io"],"resources":["projects"]},{"verbs":["get","list","watch"],"apiGroups":[""],"resources":["pods/attach","pods/exec","pods/portforward","pods/proxy","secrets","services/proxy"]},{"verbs":["impersonate"],"apiGroups":[""],"resources":["serviceaccounts"]},{"verbs":["create","delete","deletecollection","patch","update"],"apiGroups":[""],"resources":["pods","pods/attach","pods/exec","pods/portforward","pods/proxy"]},{"verbs":["create"],"apiGroups":[""],"resources":["pods/eviction"]},{"verbs":["create","delete","deletecollection","patch","update"],"apiGroups":[""],"resources":["configmaps","endpoints","events","persistentvolumeclaims","replicationcontrollers","replicationcontrollers/scale","secrets","serviceaccounts","services","services/proxy"]},{"verbs":["create"],"apiGroups":[""],"resources":["serviceaccounts/token"]},{"verbs":["create","delete","deletecollection","patch","update"],"apiGroups":["apps"],"resources":["daemonsets","deployments","deployments/rollback","deployments/scale","replicasets","replicasets/scale","statefulsets","statefulsets/scale"]},{"verbs":["create","delete","deletecollection","patch","update"],"apiGroups":["autoscaling"],"resources":["horizontalpodautoscalers"]},{"verbs":["create","delete","deletecollection","patch","update"],"apiGroups":["batch"],"resources":["cronjobs","jobs"]},{"verbs":["create","delete","deletecollection","patch","update"],"apiGroups":["extensions"],"resources":["daemonsets","deployments","deployments/rollback","deployments/scale","ingresses","networkpolicies","replicasets","replicasets/scale","replicationcontrollers/scale"]},{"verbs":["create","delete","deletecollection","patch","update"],"apiGroups":["policy"],"resources":["poddisruptionbudgets"]},{"verbs":["create","delete","deletecollection","patch","update"],"apiGroups":["networking.k8s.io"],"resources":["ingresses","networkpolicies"]},{"verbs":["create","delete","deletecollection","get","list","patch","update","watch"],"apiGroups":["coordination.k8s.io"],"resources":["leases"]},{"verbs":["create"],"apiGroups":["","image.openshift.io"],"resources":["imagestreams"]},{"verbs":["update"],"apiGroups":["","build.openshift.io"],"resources":["builds/details"]},{"verbs":["get"],"apiGroups":["","build.openshift.io"],"resources":["builds"]},{"verbs":["get","list","watch","create","update","patch","delete","deletecollection"],"apiGroups":["snapshot.storage.k8s.io"],"resources":["volumesnapshots"]},{"verbs":["create","delete","deletecollection","get","list","patch","update","watch"],"apiGroups":["","build.openshift.io"],"resources":["buildconfigs","buildconfigs/webhooks","builds"]},{"verbs":["get","list","watch"],"apiGroups":["","build.openshift.io"],"resources":["builds/log"]},{"verbs":["create"],"apiGroups":["","build.openshift.io"],"resources":["buildconfigs/instantiate","buildconfigs/instantiatebinary","builds/clone"]},{"verbs":["edit","view"],"apiGroups":["build.openshift.io"],"resources":["jenkins"]},{"verbs":["create","delete","deletecollection","get","list","patch","update","watch"],"apiGroups":["","apps.openshift.io"],"resources":["deploymentconfigs","deploymentconfigs/scale"]},{"verbs":["create"],"apiGroups":["","apps.openshift.io"],"resources":["deploymentconfigrollbacks","deploymentconfigs/instantiate","deploymentconfigs/rollback"]},{"verbs":["get","list","watch"],"apiGroups":["","apps.openshift.io"],"resources":["deploymentconfigs/log","deploymentconfigs/status"]},{"verbs":["get","list","watch"],"apiGroups":["","image.openshift.io"],"resources":["imagestreams/status"]},{"verbs":["get","list","watch"],"apiGroups":["","quota.openshift.io"],"resources":["appliedclusterresourcequotas"]},{"verbs":["create","delete","deletecollection","get","list","patch","update","watch"],"apiGroups":["","route.openshift.io"],"resources":["routes"]},{"verbs":["create"],"apiGroups":["","route.openshift.io"],"resources":["routes/custom-host"]},{"verbs":["get","list","watch"],"apiGroups":["","route.openshift.io"],"resources":["routes/status"]},{"verbs":["create","delete","deletecollection","get","list","patch","update","watch"],"apiGroups":["","template.openshift.io"],"resources":["processedtemplates","templateconfigs","templateinstances","templates"]},{"verbs":["create","delete","deletecollection","get","list","patch","update","watch"],"apiGroups":["networking.k8s.io"],"resources":["networkpolicies"]},{"verbs":["create","delete","deletecollection","get","list","patch","update","watch"],"apiGroups":["","build.openshift.io"],"resources":["buildlogs"]},{"verbs":["get","list","watch"],"apiGroups":[""],"resources":["resourcequotausages"]},{"verbs":["get","list","watch"],"apiGroups":["k8s.ovn.org"],"resources":["userdefinednetworks"]},{"verbs":["get","list","watch"],"apiGroups":["","image.openshift.io"],"resources":["imagestreamimages","imagestreammappings","imagestreams","imagestreamtags","imagetags"]},{"verbs":["get"],"apiGroups":["","image.openshift.io"],"resources":["imagestreams/layers"]},{"verbs":["get","list","watch"],"apiGroups":[""],"resources":["configmaps","endpoints","persistentvolumeclaims","persistentvolumeclaims/status","pods","replicationcontrollers","replicationcontrollers/scale","serviceaccounts","services","services/status"]},{"verbs":["get","list","watch"],"apiGroups":[""],"resources":["bindings","events","limitranges","namespaces/status","pods/log","pods/status","replicationcontrollers/status","resourcequotas","resourcequotas/status"]},{"verbs":["get","list","watch"],"apiGroups":[""],"resources":["namespaces"]},{"verbs":["get","list","watch"],"apiGroups":["discovery.k8s.io"],"resources":["endpointslices"]},{"verbs":["get","list","watch"],"apiGroups":["apps"],"resources":["controllerrevisions","daemonsets","daemonsets/status","deployments","deployments/scale","deployments/status","replicasets","replicasets/scale","replicasets/status","statefulsets","statefulsets/scale","statefulsets/status"]},{"verbs":["get","list","watch"],"apiGroups":["autoscaling"],"resources":["horizontalpodautoscalers","horizontalpodautoscalers/status"]},{"verbs":["get","list","watch"],"apiGroups":["batch"],"resources":["cronjobs","cronjobs/status","jobs","jobs/status"]},{"verbs":["get","list","watch"],"apiGroups":["extensions"],"resources":["daemonsets","daemonsets/status","deployments","deployments/scale","deployments/status","ingresses","ingresses/status","networkpolicies","replicasets","replicasets/scale","replicasets/status","replicationcontrollers/scale"]},{"verbs":["get","list","watch"],"apiGroups":["policy"],"resources":["poddisruptionbudgets","poddisruptionbudgets/status"]},{"verbs":["get","list","watch"],"apiGroups":["networking.k8s.io"],"resources":["ingresses","ingresses/status","networkpolicies"]},{"verbs":["get","list","watch"],"apiGroups":["snapshot.storage.k8s.io"],"resources":["volumesnapshots"]},{"verbs":["get","list","watch"],"apiGroups":["","build.openshift.io"],"resources":["buildconfigs","buildconfigs/webhooks","builds"]},{"verbs":["view"],"apiGroups":["build.openshift.io"],"resources":["jenkins"]},{"verbs":["get","list","watch"],"apiGroups":["","apps.openshift.io"],"resources":["deploymentconfigs","deploymentconfigs/scale"]},{"verbs":["get","list","watch"],"apiGroups":["","route.openshift.io"],"resources":["routes"]},{"verbs":["get","list","watch"],"apiGroups":["","template.openshift.io"],"resources":["processedtemplates","templateconfigs","templateinstances","templates"]},{"verbs":["get","list","watch"],"apiGroups":["","build.openshift.io"],"resources":["buildlogs"]},{"verbs":["watch","list","get"],"apiGroups":["k8s.cni.cncf.io"],"resources":["network-attachment-definitions"]},{"verbs":["create","delete","deletecollection","get","list","patch","update","watch"],"apiGroups":["","authorization.openshift.io"],"resources":["rolebindings","roles"]},{"verbs":["create","delete","deletecollection","get","list","patch","update","watch"],"apiGroups":["rbac.authorization.k8s.io"],"resources":["rolebindings","roles"]},{"verbs":["create"],"apiGroups":["","authorization.openshift.io"],"resources":["localresourceaccessreviews","localsubjectaccessreviews","subjectrulesreviews"]},{"verbs":["create"],"apiGroups":["authorization.k8s.io"],"resources":["localsubjectaccessreviews"]},{"verbs":["delete","get"],"apiGroups":["","project.openshift.io"],"resources":["projects"]},{"verbs":["create"],"apiGroups":["","authorization.openshift.io"],"resources":["resourceaccessreviews","subjectaccessreviews"]},{"verbs":["create"],"apiGroups":["","security.openshift.io"],"resources":["podsecuritypolicyreviews","podsecuritypolicyselfsubjectreviews","podsecuritypolicysubjectreviews"]},{"verbs":["get","list","watch"],"apiGroups":["","authorization.openshift.io"],"resources":["rolebindingrestrictions"]},{"verbs":["admin","edit","view"],"apiGroups":["build.openshift.io"],"resources":["jenkins"]},{"verbs":["delete","get","patch","update"],"apiGroups":["","project.openshift.io"],"resources":["projects"]},{"verbs":["update"],"apiGroups":["","route.openshift.io"],"resources":["routes/status"]}],"aggregationRule":{"clusterRoleSelectors":[{"matchLabels":{"rbac.authorization.k8s.io/aggregate-to-admin":"true"}}]}}config/validatingwebhookconfigurations/performance-addon-operator.json0000640000000000000000000000552115114274061025275 0ustar0000000000000000{"metadata":{"name":"performance-addon-operator","uid":"edb41836-67e8-444d-9434-c2163f6f7f66","resourceVersion":"4904","generation":2,"creationTimestamp":"2025-12-04T11:31:41Z","annotations":{"capability.openshift.io/name":"NodeTuning","include.release.openshift.io/hypershift":"true","include.release.openshift.io/ibm-cloud-managed":"true","include.release.openshift.io/self-managed-high-availability":"true","include.release.openshift.io/single-node-developer":"true","service.beta.openshift.io/inject-cabundle":"true"},"ownerReferences":[{"apiVersion":"config.openshift.io/v1","kind":"ClusterVersion","name":"version","uid":"a151bf9a-21b6-4d33-89ec-0e4f83e8eab6","controller":true}]},"webhooks":[{"name":"vwb.performance.openshift.io","clientConfig":{"service":{"namespace":"openshift-cluster-node-tuning-operator","name":"performance-addon-operator-service","path":"/validate-performance-openshift-io-v2-performanceprofile","port":443},"caBundle":"eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4"},"rules":[{"operations":["CREATE","UPDATE"],"apiGroups":["performance.openshift.io"],"apiVersions":["v2"],"resources":["performanceprofiles"],"scope":"*"}],"failurePolicy":"Ignore","matchPolicy":"Equivalent","namespaceSelector":{},"objectSelector":{},"sideEffects":"None","timeoutSeconds":10,"admissionReviewVersions":["v1"]}]}config/validatingwebhookconfigurations/network-node-identity.openshift.io.json0000640000000000000000000001075515114274061026734 0ustar0000000000000000{"metadata":{"name":"network-node-identity.openshift.io","uid":"bda40f06-7819-49e4-9a67-79342aae54e4","resourceVersion":"6994","generation":1,"creationTimestamp":"2025-12-04T11:38:32Z","ownerReferences":[{"apiVersion":"operator.openshift.io/v1","kind":"Network","name":"cluster","uid":"048dc3f2-5b4f-4de2-83ca-147584e743ba","controller":true,"blockOwnerDeletion":true}]},"webhooks":[{"name":"node.network-node-identity.openshift.io","clientConfig":{"url":"https://127.0.0.1:9743/node","caBundle":"eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHg="},"rules":[{"operations":["UPDATE"],"apiGroups":["*"],"apiVersions":["*"],"resources":["nodes/status"],"scope":"*"}],"failurePolicy":"Fail","matchPolicy":"Equivalent","namespaceSelector":{},"objectSelector":{},"sideEffects":"None","timeoutSeconds":10,"admissionReviewVersions":["v1"]},{"name":"pod.network-node-identity.openshift.io","clientConfig":{"url":"https://127.0.0.1:9743/pod","caBundle":"eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHg="},"rules":[{"operations":["UPDATE"],"apiGroups":["*"],"apiVersions":["*"],"resources":["pods/status"],"scope":"*"}],"failurePolicy":"Fail","matchPolicy":"Equivalent","namespaceSelector":{},"objectSelector":{},"sideEffects":"None","timeoutSeconds":10,"admissionReviewVersions":["v1"]}]}config/validatingwebhookconfigurations/multus.openshift.io.json0000640000000000000000000000506715114274061024022 0ustar0000000000000000{"metadata":{"name":"multus.openshift.io","uid":"2ff964df-b1bd-40d9-b215-b95eb5dafb51","resourceVersion":"4905","generation":2,"creationTimestamp":"2025-12-04T11:36:34Z","labels":{"app":"multus-admission-controller"},"annotations":{"service.beta.openshift.io/inject-cabundle":"true"},"ownerReferences":[{"apiVersion":"operator.openshift.io/v1","kind":"Network","name":"cluster","uid":"048dc3f2-5b4f-4de2-83ca-147584e743ba","controller":true,"blockOwnerDeletion":true}]},"webhooks":[{"name":"multus-validating-config.k8s.io","clientConfig":{"service":{"namespace":"openshift-multus","name":"multus-admission-controller","path":"/validate","port":443},"caBundle":"eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4"},"rules":[{"operations":["CREATE","UPDATE"],"apiGroups":["k8s.cni.cncf.io"],"apiVersions":["v1"],"resources":["network-attachment-definitions"],"scope":"*"}],"failurePolicy":"Fail","matchPolicy":"Equivalent","namespaceSelector":{},"objectSelector":{},"sideEffects":"NoneOnDryRun","timeoutSeconds":30,"admissionReviewVersions":["v1"]}]}config/validatingwebhookconfigurations/controlplanemachineset.machine.openshift.io.json0000640000000000000000000000537315114274061030635 0ustar0000000000000000{"metadata":{"name":"controlplanemachineset.machine.openshift.io","uid":"cad3fac6-0f85-4bd2-930d-f2c5facb022b","resourceVersion":"10943","generation":2,"creationTimestamp":"2025-12-04T11:50:30Z","annotations":{"capability.openshift.io/name":"MachineAPI","exclude.release.openshift.io/internal-openshift-hosted":"true","include.release.openshift.io/self-managed-high-availability":"true","service.beta.openshift.io/inject-cabundle":"true"},"ownerReferences":[{"apiVersion":"config.openshift.io/v1","kind":"ClusterVersion","name":"version","uid":"a151bf9a-21b6-4d33-89ec-0e4f83e8eab6","controller":true}]},"webhooks":[{"name":"controlplanemachineset.machine.openshift.io","clientConfig":{"service":{"namespace":"openshift-machine-api","name":"control-plane-machine-set-operator","path":"/validate-machine-openshift-io-v1-controlplanemachineset","port":9443},"caBundle":"eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4"},"rules":[{"operations":["CREATE","UPDATE"],"apiGroups":["machine.openshift.io"],"apiVersions":["v1"],"resources":["controlplanemachinesets"],"scope":"*"}],"failurePolicy":"Fail","matchPolicy":"Equivalent","namespaceSelector":{},"objectSelector":{},"sideEffects":"None","timeoutSeconds":10,"admissionReviewVersions":["v1"]}]}config/validatingwebhookconfigurations/cluster-baremetal-validating-webhook-configuration.json0000640000000000000000000000500215114274061032106 0ustar0000000000000000{"metadata":{"name":"cluster-baremetal-validating-webhook-configuration","uid":"0bf2ba5f-194d-40a0-9a3e-68132819160b","resourceVersion":"10793","generation":2,"creationTimestamp":"2025-12-04T11:45:55Z","annotations":{"include.release.openshift.io/self-managed-high-availability":"true","include.release.openshift.io/single-node-developer":"true","service.beta.openshift.io/inject-cabundle":"true"}},"webhooks":[{"name":"vprovisioning.kb.io","clientConfig":{"service":{"namespace":"openshift-machine-api","name":"cluster-baremetal-webhook-service","path":"/validate-metal3-io-v1alpha1-provisioning","port":443},"caBundle":"eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4"},"rules":[{"operations":["CREATE","UPDATE"],"apiGroups":["metal3.io"],"apiVersions":["v1alpha1"],"resources":["provisionings"],"scope":"*"}],"failurePolicy":"Ignore","matchPolicy":"Equivalent","namespaceSelector":{},"objectSelector":{},"sideEffects":"None","timeoutSeconds":10,"admissionReviewVersions":["v1","v1beta1"]}]}config/validatingwebhookconfigurations/autoscaling.openshift.io.json0000640000000000000000000001102715114274061024773 0ustar0000000000000000{"metadata":{"name":"autoscaling.openshift.io","uid":"c5c7c148-26ad-4fb8-9247-bbaada1b0e58","resourceVersion":"14276","generation":4,"creationTimestamp":"2025-12-04T11:46:04Z","labels":{"k8s-app":"cluster-autoscaler-operator"},"annotations":{"service.beta.openshift.io/inject-cabundle":"true"}},"webhooks":[{"name":"clusterautoscalers.autoscaling.openshift.io","clientConfig":{"service":{"namespace":"openshift-machine-api","name":"cluster-autoscaler-operator","path":"/validate-clusterautoscalers","port":443},"caBundle":"eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4"},"rules":[{"operations":["CREATE","UPDATE"],"apiGroups":["autoscaling.openshift.io"],"apiVersions":["v1"],"resources":["clusterautoscalers"],"scope":"*"}],"failurePolicy":"Ignore","matchPolicy":"Equivalent","namespaceSelector":{},"objectSelector":{},"sideEffects":"None","timeoutSeconds":10,"admissionReviewVersions":["v1"]},{"name":"machineautoscalers.autoscaling.openshift.io","clientConfig":{"service":{"namespace":"openshift-machine-api","name":"cluster-autoscaler-operator","path":"/validate-machineautoscalers","port":443},"caBundle":"eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4"},"rules":[{"operations":["CREATE","UPDATE"],"apiGroups":["autoscaling.openshift.io"],"apiVersions":["v1beta1"],"resources":["machineautoscalers"],"scope":"*"}],"failurePolicy":"Ignore","matchPolicy":"Equivalent","namespaceSelector":{},"objectSelector":{},"sideEffects":"None","timeoutSeconds":10,"admissionReviewVersions":["v1"]}]}config/clusteroperator/imageregistry.operator.openshift.io/config/cluster.json0000640000000000000000000000422315114274061026676 0ustar0000000000000000{"metadata":{"name":"cluster","uid":"dca8e85d-cea8-4ef9-bd8d-8e53b6390726","resourceVersion":"14294","generation":1,"creationTimestamp":"2025-12-04T12:00:01Z","finalizers":["imageregistry.operator.openshift.io/finalizer"]},"spec":{"managementState":"Removed","logLevel":"Normal","operatorLogLevel":"Normal","unsupportedConfigOverrides":null,"observedConfig":null,"proxy":{},"storage":{},"requests":{"read":{"maxWaitInQueue":"0s"},"write":{"maxWaitInQueue":"0s"}},"replicas":1,"rolloutStrategy":"RollingUpdate"},"status":{"observedGeneration":1,"conditions":[{"type":"ImageConfigControllerDegraded","status":"False","lastTransitionTime":"2025-12-04T12:00:01Z","reason":"AsExpected"},{"type":"AzureStackCloudControllerDegraded","status":"False","lastTransitionTime":"2025-12-04T12:00:01Z","reason":"AsExpected"},{"type":"ImageRegistryCertificatesControllerDegraded","status":"False","lastTransitionTime":"2025-12-04T12:00:02Z","reason":"AsExpected"},{"type":"Progressing","status":"False","lastTransitionTime":"2025-12-04T12:01:42Z","reason":"Removed","message":"All registry resources are removed"},{"type":"Available","status":"True","lastTransitionTime":"2025-12-04T12:00:03Z","reason":"Removed","message":"The registry is removed"},{"type":"Degraded","status":"False","lastTransitionTime":"2025-12-04T12:00:03Z","reason":"Removed","message":"The registry is removed"},{"type":"Removed","status":"True","lastTransitionTime":"2025-12-04T12:00:03Z","reason":"Removed","message":"The registry is removed"},{"type":"NodeCADaemonAvailable","status":"False","lastTransitionTime":"2025-12-04T12:01:39Z","reason":"NoAvailableReplicas","message":"The daemon set node-ca does not have available replicas"},{"type":"NodeCADaemonProgressing","status":"True","lastTransitionTime":"2025-12-04T12:01:39Z","reason":"Progressing","message":"The daemon set node-ca is updating node pods"},{"type":"NodeCADaemonControllerDegraded","status":"False","lastTransitionTime":"2025-12-04T12:01:39Z","reason":"AsExpected"}],"readyReplicas":0,"generations":[{"group":"apps","resource":"daemonsets","namespace":"openshift-image-registry","name":"node-ca","lastGeneration":1,"hash":""}],"storageManaged":false,"storage":{}}}config/apiserver.json0000640000000000000000000000121615114274061013401 0ustar0000000000000000{"metadata":{"name":"cluster","uid":"c0f9efa3-aa67-495a-9acd-bd4de32fb645","resourceVersion":"754","generation":1,"creationTimestamp":"2025-12-04T11:31:14Z","annotations":{"include.release.openshift.io/ibm-cloud-managed":"true","include.release.openshift.io/self-managed-high-availability":"true","oauth-apiserver.openshift.io/secure-token-storage":"true","release.openshift.io/create-only":"true"},"ownerReferences":[{"apiVersion":"config.openshift.io/v1","kind":"ClusterVersion","name":"version","uid":"a151bf9a-21b6-4d33-89ec-0e4f83e8eab6"}]},"spec":{"servingCerts":{},"clientCA":{"name":""},"encryption":{},"audit":{"profile":"Default"}},"status":{}}events/openshift-machine-api.json0000640000000000000000000003727015114274061015621 0ustar0000000000000000{"items":[{"namespace":"openshift-machine-api","lastTimestamp":"2025-12-04T11:39:58Z","reason":"FailedCreatePodSandBox","message":"Failed to create pod sandbox: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_machine-api-operator-88d48b57d-9fjtd_openshift-machine-api_c50317d3-f7cd-4133-845e-44add57ac378_0(a7ea781eb62388abf3d4251d464a1c215aed96cb83a2b09a854429643ac8236c): error adding pod openshift-machine-api_machine-api-operator-88d48b57d-9fjtd to CNI network \"multus-cni-network\": plugin type=\"multus-shim\" name=\"multus-cni-network\" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:\"a7ea781eb62388abf3d4251d464a1c215aed96cb83a2b09a854429643ac8236c\" Netns:\"/var/run/netns/f0df9b09-6ce7-41e5-a134-04f90f5a7bad\" IfName:\"eth0\" Args:\"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-machine-api;K8S_POD_NAME=machine-api-operator-88d48b57d-9fjtd;K8S_POD_INFRA_CONTAINER_ID=a7ea781eb62388abf3d4251d464a1c215aed96cb83a2b09a854429643ac8236c;K8S_POD_UID=c50317d3-f7cd-4133-845e-44add57ac378\" Path:\"\" ERRORED: error configuring pod [openshift-machine-api/machine-api-operator-88d48b57d-9fjtd] networking: Multus: [openshift-machine-api/machine-api-operator-88d48b57d-9fjtd/c50317d3-f7cd-4133-845e-44add57ac378]: error setting the networks status: SetPodNetworkStatusAnnotation: failed to update the pod machine-api-operator-88d48b57d-9fjtd in out of cluster comm: SetNetworkStatus: failed to update the pod machine-api-operator-88d48b57d-9fjtd in out of cluster comm: status update failed for pod /: Get \"https://api-int.sno.openstack.lab:6443/api/v1/namespaces/openshift-machine-api/pods/machine-api-operator-88d48b57d-9fjtd?timeout=1m0s\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)\n': StdinData: {\"binDir\":\"/var/lib/cni/bin\",\"clusterNetwork\":\"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf\",\"cniVersion\":\"0.3.1\",\"daemonSocketDir\":\"/run/multus/socket\",\"globalNamespaces\":\"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv\",\"logLevel\":\"verbose\",\"logToStderr\":true,\"name\":\"multus-cni-network\",\"namespaceIsolation\":true,\"type\":\"multus-shim\"}","type":"Warning"},{"namespace":"openshift-machine-api","lastTimestamp":"2025-12-04T11:41:00Z","reason":"FailedCreatePodSandBox","message":"Failed to create pod sandbox: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_machine-api-operator-88d48b57d-9fjtd_openshift-machine-api_c50317d3-f7cd-4133-845e-44add57ac378_0(af14fe68c8083f31be169d50a9e29c2a4846a7a7df4592fdfda1587a1b3dc88a): error adding pod openshift-machine-api_machine-api-operator-88d48b57d-9fjtd to CNI network \"multus-cni-network\": plugin type=\"multus-shim\" name=\"multus-cni-network\" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:\"af14fe68c8083f31be169d50a9e29c2a4846a7a7df4592fdfda1587a1b3dc88a\" Netns:\"/var/run/netns/d8d0e1e9-6bc3-435a-b27b-ff4ea6b4043d\" IfName:\"eth0\" Args:\"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-machine-api;K8S_POD_NAME=machine-api-operator-88d48b57d-9fjtd;K8S_POD_INFRA_CONTAINER_ID=af14fe68c8083f31be169d50a9e29c2a4846a7a7df4592fdfda1587a1b3dc88a;K8S_POD_UID=c50317d3-f7cd-4133-845e-44add57ac378\" Path:\"\" ERRORED: error configuring pod [openshift-machine-api/machine-api-operator-88d48b57d-9fjtd] networking: Multus: [openshift-machine-api/machine-api-operator-88d48b57d-9fjtd/c50317d3-f7cd-4133-845e-44add57ac378]: error setting the networks status: SetPodNetworkStatusAnnotation: failed to update the pod machine-api-operator-88d48b57d-9fjtd in out of cluster comm: SetNetworkStatus: failed to update the pod machine-api-operator-88d48b57d-9fjtd in out of cluster comm: status update failed for pod /: Get \"https://api-int.sno.openstack.lab:6443/api/v1/namespaces/openshift-machine-api/pods/machine-api-operator-88d48b57d-9fjtd?timeout=1m0s\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)\n': StdinData: {\"binDir\":\"/var/lib/cni/bin\",\"clusterNetwork\":\"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf\",\"cniVersion\":\"0.3.1\",\"daemonSocketDir\":\"/run/multus/socket\",\"globalNamespaces\":\"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv\",\"logLevel\":\"verbose\",\"logToStderr\":true,\"name\":\"multus-cni-network\",\"namespaceIsolation\":true,\"type\":\"multus-shim\"}","type":"Warning"},{"namespace":"openshift-machine-api","lastTimestamp":"2025-12-04T11:42:01Z","reason":"FailedCreatePodSandBox","message":"Failed to create pod sandbox: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_machine-api-operator-88d48b57d-9fjtd_openshift-machine-api_c50317d3-f7cd-4133-845e-44add57ac378_0(21bc0e2f8c3e2b30320dbc8a79d7d97b34ea191a5786a54f69ac61586aa362d5): error adding pod openshift-machine-api_machine-api-operator-88d48b57d-9fjtd to CNI network \"multus-cni-network\": plugin type=\"multus-shim\" name=\"multus-cni-network\" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:\"21bc0e2f8c3e2b30320dbc8a79d7d97b34ea191a5786a54f69ac61586aa362d5\" Netns:\"/var/run/netns/2052808d-2414-40cf-bf30-07beb222e23d\" IfName:\"eth0\" Args:\"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-machine-api;K8S_POD_NAME=machine-api-operator-88d48b57d-9fjtd;K8S_POD_INFRA_CONTAINER_ID=21bc0e2f8c3e2b30320dbc8a79d7d97b34ea191a5786a54f69ac61586aa362d5;K8S_POD_UID=c50317d3-f7cd-4133-845e-44add57ac378\" Path:\"\" ERRORED: error configuring pod [openshift-machine-api/machine-api-operator-88d48b57d-9fjtd] networking: Multus: [openshift-machine-api/machine-api-operator-88d48b57d-9fjtd/c50317d3-f7cd-4133-845e-44add57ac378]: error setting the networks status: SetPodNetworkStatusAnnotation: failed to update the pod machine-api-operator-88d48b57d-9fjtd in out of cluster comm: SetNetworkStatus: failed to update the pod machine-api-operator-88d48b57d-9fjtd in out of cluster comm: status update failed for pod /: Get \"https://api-int.sno.openstack.lab:6443/api/v1/namespaces/openshift-machine-api/pods/machine-api-operator-88d48b57d-9fjtd?timeout=1m0s\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)\n': StdinData: {\"binDir\":\"/var/lib/cni/bin\",\"clusterNetwork\":\"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf\",\"cniVersion\":\"0.3.1\",\"daemonSocketDir\":\"/run/multus/socket\",\"globalNamespaces\":\"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv\",\"logLevel\":\"verbose\",\"logToStderr\":true,\"name\":\"multus-cni-network\",\"namespaceIsolation\":true,\"type\":\"multus-shim\"}","type":"Warning"},{"namespace":"openshift-machine-api","lastTimestamp":"2025-12-04T11:43:03Z","reason":"FailedCreatePodSandBox","message":"Failed to create pod sandbox: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_machine-api-operator-88d48b57d-9fjtd_openshift-machine-api_c50317d3-f7cd-4133-845e-44add57ac378_0(97bd237a7b73814f3563223ef08fe64c059bb01e4f8387a182eaaae722ea06a1): error adding pod openshift-machine-api_machine-api-operator-88d48b57d-9fjtd to CNI network \"multus-cni-network\": plugin type=\"multus-shim\" name=\"multus-cni-network\" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:\"97bd237a7b73814f3563223ef08fe64c059bb01e4f8387a182eaaae722ea06a1\" Netns:\"/var/run/netns/577c5238-fde8-4e8e-9202-5e640ae52ce7\" IfName:\"eth0\" Args:\"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-machine-api;K8S_POD_NAME=machine-api-operator-88d48b57d-9fjtd;K8S_POD_INFRA_CONTAINER_ID=97bd237a7b73814f3563223ef08fe64c059bb01e4f8387a182eaaae722ea06a1;K8S_POD_UID=c50317d3-f7cd-4133-845e-44add57ac378\" Path:\"\" ERRORED: error configuring pod [openshift-machine-api/machine-api-operator-88d48b57d-9fjtd] networking: Multus: [openshift-machine-api/machine-api-operator-88d48b57d-9fjtd/c50317d3-f7cd-4133-845e-44add57ac378]: error setting the networks status: SetPodNetworkStatusAnnotation: failed to update the pod machine-api-operator-88d48b57d-9fjtd in out of cluster comm: SetNetworkStatus: failed to update the pod machine-api-operator-88d48b57d-9fjtd in out of cluster comm: status update failed for pod /: Get \"https://api-int.sno.openstack.lab:6443/api/v1/namespaces/openshift-machine-api/pods/machine-api-operator-88d48b57d-9fjtd?timeout=1m0s\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)\n': StdinData: {\"binDir\":\"/var/lib/cni/bin\",\"clusterNetwork\":\"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf\",\"cniVersion\":\"0.3.1\",\"daemonSocketDir\":\"/run/multus/socket\",\"globalNamespaces\":\"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv\",\"logLevel\":\"verbose\",\"logToStderr\":true,\"name\":\"multus-cni-network\",\"namespaceIsolation\":true,\"type\":\"multus-shim\"}","type":"Warning"},{"namespace":"openshift-machine-api","lastTimestamp":"2025-12-04T11:44:04Z","reason":"FailedCreatePodSandBox","message":"Failed to create pod sandbox: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_machine-api-operator-88d48b57d-9fjtd_openshift-machine-api_c50317d3-f7cd-4133-845e-44add57ac378_0(b8a357e4d9b7049620c368e3117a218cab578e2787de9134003b973d9b4b0f47): error adding pod openshift-machine-api_machine-api-operator-88d48b57d-9fjtd to CNI network \"multus-cni-network\": plugin type=\"multus-shim\" name=\"multus-cni-network\" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:\"b8a357e4d9b7049620c368e3117a218cab578e2787de9134003b973d9b4b0f47\" Netns:\"/var/run/netns/afdfd7cd-85b7-4aec-840d-af54b181ee50\" IfName:\"eth0\" Args:\"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-machine-api;K8S_POD_NAME=machine-api-operator-88d48b57d-9fjtd;K8S_POD_INFRA_CONTAINER_ID=b8a357e4d9b7049620c368e3117a218cab578e2787de9134003b973d9b4b0f47;K8S_POD_UID=c50317d3-f7cd-4133-845e-44add57ac378\" Path:\"\" ERRORED: error configuring pod [openshift-machine-api/machine-api-operator-88d48b57d-9fjtd] networking: Multus: [openshift-machine-api/machine-api-operator-88d48b57d-9fjtd/c50317d3-f7cd-4133-845e-44add57ac378]: error setting the networks status: SetPodNetworkStatusAnnotation: failed to update the pod machine-api-operator-88d48b57d-9fjtd in out of cluster comm: SetNetworkStatus: failed to update the pod machine-api-operator-88d48b57d-9fjtd in out of cluster comm: status update failed for pod /: Get \"https://api-int.sno.openstack.lab:6443/api/v1/namespaces/openshift-machine-api/pods/machine-api-operator-88d48b57d-9fjtd?timeout=1m0s\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)\n': StdinData: {\"binDir\":\"/var/lib/cni/bin\",\"clusterNetwork\":\"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf\",\"cniVersion\":\"0.3.1\",\"daemonSocketDir\":\"/run/multus/socket\",\"globalNamespaces\":\"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv\",\"logLevel\":\"verbose\",\"logToStderr\":true,\"name\":\"multus-cni-network\",\"namespaceIsolation\":true,\"type\":\"multus-shim\"}","type":"Warning"},{"namespace":"openshift-machine-api","lastTimestamp":"2025-12-04T11:45:05Z","reason":"FailedCreatePodSandBox","message":"Failed to create pod sandbox: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_machine-api-operator-88d48b57d-9fjtd_openshift-machine-api_c50317d3-f7cd-4133-845e-44add57ac378_0(5d2384d86be2f95b3966a53aba3cff3d8ac082133332b4d7e800c63a2574036a): error adding pod openshift-machine-api_machine-api-operator-88d48b57d-9fjtd to CNI network \"multus-cni-network\": plugin type=\"multus-shim\" name=\"multus-cni-network\" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:\"5d2384d86be2f95b3966a53aba3cff3d8ac082133332b4d7e800c63a2574036a\" Netns:\"/var/run/netns/5587937f-ddb7-4990-8879-cff0c5f010ae\" IfName:\"eth0\" Args:\"IgnoreUnknown=1;K8S_POD_NAMESPACE=openshift-machine-api;K8S_POD_NAME=machine-api-operator-88d48b57d-9fjtd;K8S_POD_INFRA_CONTAINER_ID=5d2384d86be2f95b3966a53aba3cff3d8ac082133332b4d7e800c63a2574036a;K8S_POD_UID=c50317d3-f7cd-4133-845e-44add57ac378\" Path:\"\" ERRORED: error configuring pod [openshift-machine-api/machine-api-operator-88d48b57d-9fjtd] networking: Multus: [openshift-machine-api/machine-api-operator-88d48b57d-9fjtd/c50317d3-f7cd-4133-845e-44add57ac378]: error setting the networks status: SetPodNetworkStatusAnnotation: failed to update the pod machine-api-operator-88d48b57d-9fjtd in out of cluster comm: SetNetworkStatus: failed to update the pod machine-api-operator-88d48b57d-9fjtd in out of cluster comm: status update failed for pod /: Get \"https://api-int.sno.openstack.lab:6443/api/v1/namespaces/openshift-machine-api/pods/machine-api-operator-88d48b57d-9fjtd?timeout=1m0s\": context deadline exceeded\n': StdinData: {\"binDir\":\"/var/lib/cni/bin\",\"clusterNetwork\":\"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf\",\"cniVersion\":\"0.3.1\",\"daemonSocketDir\":\"/run/multus/socket\",\"globalNamespaces\":\"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv\",\"logLevel\":\"verbose\",\"logToStderr\":true,\"name\":\"multus-cni-network\",\"namespaceIsolation\":true,\"type\":\"multus-shim\"}","type":"Warning"},{"namespace":"openshift-machine-api","lastTimestamp":"2025-12-04T11:54:06Z","reason":"BackOff","message":"Back-off restarting failed container cluster-baremetal-operator in pod cluster-baremetal-operator-78f758c7b9-zgkh5_openshift-machine-api(7e3160a9-11d1-4845-ba30-1a49ae7339a9)","type":"Warning"},{"namespace":"openshift-machine-api","lastTimestamp":"2025-12-04T11:55:49Z","reason":"BackOff","message":"Back-off restarting failed container cluster-autoscaler-operator in pod cluster-autoscaler-operator-5f49d774cd-894dk_openshift-machine-api(e7fc7c16-5bca-49e5-aff0-7a8f80c6b639)","type":"Warning"},{"namespace":"openshift-machine-api","lastTimestamp":"2025-12-04T11:59:27Z","reason":"FailedMount","message":"MountVolume.SetUp failed for volume \"config\" : failed to sync configmap cache: timed out waiting for the condition","type":"Warning"},{"namespace":"openshift-machine-api","lastTimestamp":"2025-12-04T11:59:27Z","reason":"FailedMount","message":"MountVolume.SetUp failed for volume \"images\" : failed to sync configmap cache: timed out waiting for the condition","type":"Warning"},{"namespace":"openshift-machine-api","lastTimestamp":"2025-12-04T11:59:27Z","reason":"FailedMount","message":"MountVolume.SetUp failed for volume \"cluster-baremetal-operator-tls\" : failed to sync secret cache: timed out waiting for the condition","type":"Warning"},{"namespace":"openshift-machine-api","lastTimestamp":"2025-12-04T11:59:27Z","reason":"FailedMount","message":"MountVolume.SetUp failed for volume \"cert\" : failed to sync secret cache: timed out waiting for the condition","type":"Warning"},{"namespace":"openshift-machine-api","lastTimestamp":"2025-12-04T11:59:27Z","reason":"FailedMount","message":"MountVolume.SetUp failed for volume \"cert\" : failed to sync secret cache: timed out waiting for the condition","type":"Warning"},{"namespace":"openshift-machine-api","lastTimestamp":"2025-12-04T11:59:27Z","reason":"FailedMount","message":"MountVolume.SetUp failed for volume \"auth-proxy-config\" : failed to sync configmap cache: timed out waiting for the condition","type":"Warning"},{"namespace":"openshift-machine-api","lastTimestamp":"2025-12-04T11:59:27Z","reason":"FailedMount","message":"MountVolume.SetUp failed for volume \"machine-api-operator-tls\" : failed to sync secret cache: timed out waiting for the condition","type":"Warning"},{"namespace":"openshift-machine-api","lastTimestamp":"2025-12-04T11:59:27Z","reason":"FailedMount","message":"MountVolume.SetUp failed for volume \"images\" : failed to sync configmap cache: timed out waiting for the condition","type":"Warning"},{"namespace":"openshift-machine-api","lastTimestamp":"2025-12-04T11:59:27Z","reason":"FailedMount","message":"MountVolume.SetUp failed for volume \"config\" : failed to sync configmap cache: timed out waiting for the condition","type":"Warning"}]}config/authentication.json0000640000000000000000000000132415114274061014420 0ustar0000000000000000{"metadata":{"name":"cluster","uid":"38a7f697-7219-438f-b929-98ec151d96dd","resourceVersion":"11355","generation":2,"creationTimestamp":"2025-12-04T11:31:16Z","annotations":{"include.release.openshift.io/ibm-cloud-managed":"true","include.release.openshift.io/self-managed-high-availability":"true","release.openshift.io/create-only":"true"},"ownerReferences":[{"apiVersion":"config.openshift.io/v1","kind":"ClusterVersion","name":"version","uid":"a151bf9a-21b6-4d33-89ec-0e4f83e8eab6"}]},"spec":{"type":"","oauthMetadata":{"name":""},"webhookTokenAuthenticator":{"kubeConfig":{"name":"webhook-authentication-integrated-oauth"}},"serviceAccountIssuer":""},"status":{"integratedOAuthMetadata":{"name":""},"oidcClients":null}}config/oauth.json0000640000000000000000000000114415114274061012521 0ustar0000000000000000{"metadata":{"name":"cluster","uid":"0426ec35-6688-4e4c-933c-89ef2cfbef60","resourceVersion":"1535","generation":1,"creationTimestamp":"2025-12-04T11:31:30Z","annotations":{"include.release.openshift.io/ibm-cloud-managed":"true","include.release.openshift.io/self-managed-high-availability":"true","release.openshift.io/create-only":"true"},"ownerReferences":[{"apiVersion":"config.openshift.io/v1","kind":"ClusterVersion","name":"version","uid":"a151bf9a-21b6-4d33-89ec-0e4f83e8eab6"}]},"spec":{"tokenConfig":{},"templates":{"login":{"name":""},"providerSelection":{"name":""},"error":{"name":""}}},"status":{}}config/mutatingwebhookconfigurations/catalogd-mutating-webhook-configuration.json0000640000000000000000000000533615114274061027477 0ustar0000000000000000{"metadata":{"name":"catalogd-mutating-webhook-configuration","uid":"443f4b4a-bbda-4f16-b0fe-ba7429ec8b34","resourceVersion":"5725","generation":2,"creationTimestamp":"2025-12-04T11:38:02Z","annotations":{"service.beta.openshift.io/inject-cabundle":"true"}},"webhooks":[{"name":"inject-metadata-name.olm.operatorframework.io","clientConfig":{"service":{"namespace":"openshift-catalogd","name":"catalogd-service","path":"/mutate-olm-operatorframework-io-v1-clustercatalog","port":9443},"caBundle":"eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4eHh4"},"rules":[{"operations":["CREATE","UPDATE"],"apiGroups":["olm.operatorframework.io"],"apiVersions":["v1"],"resources":["clustercatalogs"],"scope":"*"}],"failurePolicy":"Fail","matchPolicy":"Equivalent","namespaceSelector":{},"objectSelector":{},"sideEffects":"None","timeoutSeconds":10,"admissionReviewVersions":["v1"],"reinvocationPolicy":"Never","matchConditions":[{"name":"MissingOrIncorrectMetadataNameLabel","expression":"'name' in object.metadata \u0026\u0026 (!has(object.metadata.labels) || !('olm.operatorframework.io/metadata.name' in object.metadata.labels) || object.metadata.labels['olm.operatorframework.io/metadata.name'] != object.metadata.name)"}]}]}config/node/master-0.json0000640000000000000000000000551515114274061013764 0ustar0000000000000000{"metadata":{"name":"master-0","uid":"9618c86a-8a87-4c00-8dd2-9747684d64b3","resourceVersion":"14296","creationTimestamp":"2025-12-04T11:35:48Z","labels":{"beta.kubernetes.io/arch":"amd64","beta.kubernetes.io/os":"linux","kubernetes.io/arch":"amd64","kubernetes.io/hostname":"master-0","kubernetes.io/os":"linux","node-role.kubernetes.io/control-plane":"","node-role.kubernetes.io/master":"","node.openshift.io/os_id":"rhcos"},"annotations":{"k8s.ovn.org/host-cidrs":"","k8s.ovn.org/l3-gateway-config":"","k8s.ovn.org/node-chassis-id":"","k8s.ovn.org/node-encap-ips":"","k8s.ovn.org/node-id":"","k8s.ovn.org/node-masquerade-subnet":"","k8s.ovn.org/node-primary-ifaddr":"","k8s.ovn.org/node-subnets":"","k8s.ovn.org/node-transit-switch-port-ifaddr":"","k8s.ovn.org/remote-zone-migrated":"","k8s.ovn.org/zone-name":"","volumes.kubernetes.io/controller-managed-attach-detach":"true"}},"spec":{"taints":[{"key":"node-role.kubernetes.io/master","effect":"NoSchedule"}]},"status":{"capacity":{"cpu":"16","ephemeral-storage":"209124332Ki","hugepages-1Gi":"0","hugepages-2Mi":"0","memory":"49330228Ki","pods":"250"},"allocatable":{"cpu":"15500m","ephemeral-storage":"191655242229","hugepages-1Gi":"0","hugepages-2Mi":"0","memory":"48179252Ki","pods":"250"},"conditions":[{"type":"MemoryPressure","status":"False","lastHeartbeatTime":"2025-12-04T12:01:43Z","lastTransitionTime":"2025-12-04T11:35:48Z","reason":"KubeletHasSufficientMemory","message":"kubelet has sufficient memory available"},{"type":"DiskPressure","status":"False","lastHeartbeatTime":"2025-12-04T12:01:43Z","lastTransitionTime":"2025-12-04T11:35:48Z","reason":"KubeletHasNoDiskPressure","message":"kubelet has no disk pressure"},{"type":"PIDPressure","status":"False","lastHeartbeatTime":"2025-12-04T12:01:43Z","lastTransitionTime":"2025-12-04T11:35:48Z","reason":"KubeletHasSufficientPID","message":"kubelet has sufficient PID available"},{"type":"Ready","status":"True","lastHeartbeatTime":"2025-12-04T12:01:43Z","lastTransitionTime":"2025-12-04T11:37:30Z","reason":"KubeletReady","message":"kubelet is posting ready status"}],"addresses":[{"type":"InternalIP","address":"192.168.32.10"},{"type":"Hostname","address":"master-0"}],"daemonEndpoints":{"kubeletEndpoint":{"Port":10250}},"nodeInfo":{"machineID":"xxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxx","systemUUID":"xxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxx","bootID":"xxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxx","kernelVersion":"5.14.0-427.100.1.el9_4.x86_64","osImage":"Red Hat Enterprise Linux CoreOS 418.94.202511170715-0","containerRuntimeVersion":"cri-o://1.31.13-2.rhaos4.18.git15789b8.el9","kubeletVersion":"v1.31.13","kubeProxyVersion":"v1.31.13","operatingSystem":"linux","architecture":"amd64"},"runtimeHandlers":[{"name":"crun","features":{"recursiveReadOnlyMounts":true}},{"name":"","features":{"recursiveReadOnlyMounts":true}},{"name":"runc","features":{"recursiveReadOnlyMounts":true}}]}}config/featuregate.json0000640000000000000000000000613615114274061013703 0ustar0000000000000000{"metadata":{"name":"cluster","uid":"f773262d-0714-4d19-84db-d218e26433b1","resourceVersion":"641","generation":1,"creationTimestamp":"2025-12-04T11:30:55Z","annotations":{"include.release.openshift.io/self-managed-high-availability":"true"}},"spec":{},"status":{"featureGates":[{"version":"4.18.29","enabled":[{"name":"AWSEFSDriverVolumeMetrics"},{"name":"AdminNetworkPolicy"},{"name":"AlibabaPlatform"},{"name":"AzureWorkloadIdentity"},{"name":"BareMetalLoadBalancer"},{"name":"BuildCSIVolumes"},{"name":"ChunkSizeMiB"},{"name":"CloudDualStackNodeIPs"},{"name":"DisableKubeletCloudCredentialProviders"},{"name":"GCPLabelsTags"},{"name":"HardwareSpeed"},{"name":"IngressControllerLBSubnetsAWS"},{"name":"KMSv1"},{"name":"ManagedBootImages"},{"name":"ManagedBootImagesAWS"},{"name":"MultiArchInstallAWS"},{"name":"MultiArchInstallGCP"},{"name":"NetworkDiagnosticsConfig"},{"name":"NetworkLiveMigration"},{"name":"NetworkSegmentation"},{"name":"NewOLM"},{"name":"NodeDisruptionPolicy"},{"name":"OnClusterBuild"},{"name":"PersistentIPsForVirtualization"},{"name":"PrivateHostedZoneAWS"},{"name":"SetEIPForNLBIngressController"},{"name":"VSphereControlPlaneMachineSet"},{"name":"VSphereDriverConfiguration"},{"name":"VSphereMultiVCenters"},{"name":"VSphereStaticIPs"},{"name":"ValidatingAdmissionPolicy"}],"disabled":[{"name":"AWSClusterHostedDNS"},{"name":"AdditionalRoutingCapabilities"},{"name":"AutomatedEtcdBackup"},{"name":"BootcNodeManagement"},{"name":"CSIDriverSharedResource"},{"name":"ClusterAPIInstall"},{"name":"ClusterAPIInstallIBMCloud"},{"name":"ClusterMonitoringConfig"},{"name":"ConsolePluginContentSecurityPolicy"},{"name":"DNSNameResolver"},{"name":"DynamicResourceAllocation"},{"name":"EtcdBackendQuota"},{"name":"EventedPLEG"},{"name":"Example"},{"name":"ExternalOIDC"},{"name":"ExternalOIDCWithUIDAndExtraClaimMappings"},{"name":"GCPClusterHostedDNS"},{"name":"GatewayAPI"},{"name":"ImageStreamImportMode"},{"name":"IngressControllerDynamicConfigurationManager"},{"name":"InsightsConfig"},{"name":"InsightsConfigAPI"},{"name":"InsightsOnDemandDataGather"},{"name":"InsightsRuntimeExtractor"},{"name":"MachineAPIMigration"},{"name":"MachineAPIOperatorDisableMachineHealthCheckController"},{"name":"MachineAPIProviderOpenStack"},{"name":"MachineConfigNodes"},{"name":"MaxUnavailableStatefulSet"},{"name":"MetricsCollectionProfiles"},{"name":"MinimumKubeletVersion"},{"name":"MixedCPUsAllocation"},{"name":"MultiArchInstallAzure"},{"name":"NodeSwap"},{"name":"NutanixMultiSubnets"},{"name":"OVNObservability"},{"name":"OpenShiftPodSecurityAdmission"},{"name":"PinnedImages"},{"name":"PlatformOperators"},{"name":"ProcMountType"},{"name":"RouteAdvertisements"},{"name":"RouteExternalCertificate"},{"name":"ServiceAccountTokenNodeBinding"},{"name":"SignatureStores"},{"name":"SigstoreImageVerification"},{"name":"StreamingCollectionEncodingToJSON"},{"name":"StreamingCollectionEncodingToProtobuf"},{"name":"TranslateStreamCloseWebsocketRequests"},{"name":"UpgradeStatus"},{"name":"UserNamespacesPodSecurityStandards"},{"name":"UserNamespacesSupport"},{"name":"VSphereMultiNetworks"},{"name":"VolumeAttributesClass"},{"name":"VolumeGroupSnapshot"}]}]}}config/network.json0000640000000000000000000000117015114274061013071 0ustar0000000000000000{"metadata":{"name":"cluster","uid":"4d45be9c-d3dd-4f70-8855-3b691962ef3a","resourceVersion":"3219","generation":2,"creationTimestamp":"2025-12-04T11:30:34Z"},"spec":{"clusterNetwork":[{"cidr":"10.128.0.0/16","hostPrefix":23}],"serviceNetwork":["172.30.0.0/16"],"networkType":"OVNKubernetes","externalIP":{"policy":{}},"networkDiagnostics":{"mode":"","sourcePlacement":{"nodeSelector":null,"tolerations":null},"targetPlacement":{"nodeSelector":null,"tolerations":null}}},"status":{"clusterNetwork":[{"cidr":"10.128.0.0/16","hostPrefix":23}],"serviceNetwork":["172.30.0.0/16"],"networkType":"OVNKubernetes","clusterNetworkMTU":8900}}config/pdbs/openshift-operator-lifecycle-manager/packageserver-pdb.json0000640000000000000000000000174015114274061025115 0ustar0000000000000000{"metadata":{"name":"packageserver-pdb","namespace":"openshift-operator-lifecycle-manager","uid":"d2435e62-c9e5-4695-a493-ed1fdc3f8da3","resourceVersion":"8977","generation":1,"creationTimestamp":"2025-12-04T11:31:32Z","annotations":{"capability.openshift.io/name":"OperatorLifecycleManager","include.release.openshift.io/hypershift":"true","include.release.openshift.io/ibm-cloud-managed":"true","include.release.openshift.io/self-managed-high-availability":"true"},"ownerReferences":[{"apiVersion":"config.openshift.io/v1","kind":"ClusterVersion","name":"version","uid":"a151bf9a-21b6-4d33-89ec-0e4f83e8eab6","controller":true}]},"spec":{"selector":{"matchLabels":{"app":"packageserver"}},"maxUnavailable":1},"status":{"observedGeneration":1,"disruptionsAllowed":0,"currentHealthy":0,"desiredHealthy":0,"expectedPods":1,"conditions":[{"type":"DisruptionAllowed","status":"False","observedGeneration":1,"lastTransitionTime":"2025-12-04T11:31:32Z","reason":"InsufficientPods","message":""}]}}config/machineconfigpools/worker.json0000640000000000000000000000106515114274061016563 0ustar0000000000000000{"apiVersion":"machineconfiguration.openshift.io/v1","kind":"MachineConfigPool","metadata":{"creationTimestamp":"2025-12-04T11:45:11Z","generation":1,"labels":{"machineconfiguration.openshift.io/mco-built-in":"","pools.operator.machineconfiguration.openshift.io/worker":""},"name":"worker","resourceVersion":"9795","uid":"7cf320ef-501d-4439-97d1-12660d132684"},"spec":{"configuration":{},"machineConfigSelector":{"matchLabels":{"machineconfiguration.openshift.io/role":"worker"}},"nodeSelector":{"matchLabels":{"node-role.kubernetes.io/worker":""}},"paused":false}}config/machineconfigpools/master.json0000640000000000000000000000117215114274061016544 0ustar0000000000000000{"apiVersion":"machineconfiguration.openshift.io/v1","kind":"MachineConfigPool","metadata":{"creationTimestamp":"2025-12-04T11:45:11Z","generation":1,"labels":{"machineconfiguration.openshift.io/mco-built-in":"","operator.machineconfiguration.openshift.io/required-for-upgrade":"","pools.operator.machineconfiguration.openshift.io/master":""},"name":"master","resourceVersion":"9793","uid":"a9a5d593-27f8-478a-b013-b81fdd6d3e4c"},"spec":{"configuration":{},"machineConfigSelector":{"matchLabels":{"machineconfiguration.openshift.io/role":"master"}},"nodeSelector":{"matchLabels":{"node-role.kubernetes.io/master":""}},"paused":false}}config/image.json0000640000000000000000000000111715114274061012463 0ustar0000000000000000{"metadata":{"name":"cluster","uid":"68a8f50d-7c26-4828-b723-6a9984c515c7","resourceVersion":"1322","generation":1,"creationTimestamp":"2025-12-04T11:31:26Z","annotations":{"include.release.openshift.io/ibm-cloud-managed":"true","include.release.openshift.io/self-managed-high-availability":"true","release.openshift.io/create-only":"true"},"ownerReferences":[{"apiVersion":"config.openshift.io/v1","kind":"ClusterVersion","name":"version","uid":"a151bf9a-21b6-4d33-89ec-0e4f83e8eab6"}]},"spec":{"additionalTrustedCA":{"name":""},"registrySources":{},"imageStreamImportMode":""},"status":{}}config/machinehealthchecks/openshift-machine-api/machine-api-termination-handler.json0000640000000000000000000000163315114274061027644 0ustar0000000000000000{"apiVersion":"machine.openshift.io/v1beta1","kind":"MachineHealthCheck","metadata":{"annotations":{"capability.openshift.io/name":"MachineAPI","exclude.release.openshift.io/internal-openshift-hosted":"true","include.release.openshift.io/self-managed-high-availability":"true"},"creationTimestamp":"2025-12-04T11:50:40Z","generation":1,"labels":{"api":"clusterapi","k8s-app":"termination-handler"},"name":"machine-api-termination-handler","namespace":"openshift-machine-api","ownerReferences":[{"apiVersion":"config.openshift.io/v1","controller":true,"kind":"ClusterVersion","name":"version","uid":"a151bf9a-21b6-4d33-89ec-0e4f83e8eab6"}],"resourceVersion":"11003","uid":"8311a025-c4ca-47cf-aae4-21b37123fd02"},"spec":{"maxUnhealthy":"100%","nodeStartupTimeout":"0","selector":{"matchLabels":{"machine.openshift.io/interruptible-instance":""}},"unhealthyConditions":[{"status":"True","timeout":"0s","type":"Terminating"}]}}config/schedulers/cluster.json0000640000000000000000000000052415114274061015224 0ustar0000000000000000{"kind":"Scheduler","apiVersion":"config.openshift.io/v1","metadata":{"name":"cluster","uid":"9e94a158-aa27-4464-a40c-7ce9766d02f6","resourceVersion":"556","generation":1,"creationTimestamp":"2025-12-04T11:30:36Z"},"spec":{"policy":{"name":""},"profileCustomizations":{"dynamicResourceAllocation":""},"mastersSchedulable":true},"status":{}}config/clusteroperator/imageregistry.operator.openshift.io/imagepruner/cluster.json0000640000000000000000000000163115114274061027747 0ustar0000000000000000{"metadata":{"name":"cluster","uid":"fc42da9c-dfd0-4be6-8f09-0b44eeecb5ee","resourceVersion":"6425","generation":1,"creationTimestamp":"2025-12-04T11:38:19Z"},"spec":{"schedule":"","suspend":false,"keepTagRevisions":3,"successfulJobsHistoryLimit":3,"failedJobsHistoryLimit":3,"ignoreInvalidImageReferences":true,"logLevel":"Normal"},"status":{"observedGeneration":1,"conditions":[{"type":"Available","status":"True","lastTransitionTime":"2025-12-04T11:38:20Z","reason":"AsExpected","message":"Pruner CronJob has been created"},{"type":"Failed","status":"False","lastTransitionTime":"2025-12-04T11:38:20Z","reason":"Complete","message":"Pruner completed successfully"},{"type":"Scheduled","status":"True","lastTransitionTime":"2025-12-04T11:38:20Z","reason":"Scheduled","message":"The pruner job has been scheduled"},{"type":"Degraded","status":"False","lastTransitionTime":"2025-12-04T11:38:20Z","reason":"AsExpected"}]}}config/proxy.json0000640000000000000000000000031315114274061012557 0ustar0000000000000000{"metadata":{"name":"cluster","uid":"9b8ea29c-6e1e-4655-8f43-c1f5667ac5bf","resourceVersion":"551","generation":1,"creationTimestamp":"2025-12-04T11:30:35Z"},"spec":{"trustedCA":{"name":""}},"status":{}}config/infrastructure.json0000640000000000000000000000106515114274061014463 0ustar0000000000000000{"metadata":{"name":"cluster","uid":"03802ff5-a5d6-41de-854c-d993c32db0ce","resourceVersion":"533","generation":1,"creationTimestamp":"2025-12-04T11:30:33Z"},"spec":{"cloudConfig":{"name":""},"platformSpec":{"type":"None"}},"status":{"infrastructureName":"xxx-xxxxx","platform":"None","platformStatus":{"type":"None"},"etcdDiscoveryDomain":"","apiServerURL":"https://api.sno.openstack.lab:6443","apiServerInternalURI":"https://api-int.sno.openstack.lab:6443","controlPlaneTopology":"SingleReplica","infrastructureTopology":"SingleReplica","cpuPartitioning":"None"}}config/ingress.json0000640000000000000000000000213715114274061013056 0ustar0000000000000000{"metadata":{"name":"cluster","uid":"5c58a62d-96ff-4ad6-866e-43117a7d920b","resourceVersion":"12099","generation":1,"creationTimestamp":"2025-12-04T11:30:34Z"},"spec":{"domain":"apps.sno.openstack.lab","loadBalancer":{"platform":{"type":""}}},"status":{"componentRoutes":[{"namespace":"openshift-authentication","name":"oauth-openshift","defaultHostname":"oauth-openshift.apps.sno.openstack.lab","consumingUsers":["system:serviceaccount:oauth-openshift:authentication-operator"],"currentHostnames":["oauth-openshift.apps.sno.openstack.lab"],"conditions":[{"type":"Progressing","status":"True","lastTransitionTime":"2025-12-04T11:59:53Z","reason":"RouteNotAdmitted","message":"Route not admitted: no ingress for host oauth-openshift.apps.sno.openstack.lab in route oauth-openshift in namespace openshift-authentication"},{"type":"Degraded","status":"False","lastTransitionTime":"2025-12-04T11:59:53Z","reason":"AsExpected","message":"All is well"}],"relatedObjects":[{"group":"route.openshift.io","resource":"routes","namespace":"openshift-authentication","name":"oauth-openshift"}]}],"defaultPlacement":"ControlPlane"}}